public inbox for intel-gfx@lists.freedesktop.org
 help / color / mirror / Atom feed
* [PATCH] drm: Paper over locking inversion after registration rework
@ 2016-08-04 10:15 Daniel Vetter
  2016-08-04 10:35 ` Jani Nikula
                   ` (2 more replies)
  0 siblings, 3 replies; 7+ messages in thread
From: Daniel Vetter @ 2016-08-04 10:15 UTC (permalink / raw)
  To: DRI Development; +Cc: Daniel Vetter, Daniel Vetter, Intel Graphics Development

drm_connector_register_all requires a few too many locks because our
connector_list locking is busted. Add another FIXME+hack to work
around this. This should address the below lockdep splat:

======================================================
[ INFO: possible circular locking dependency detected ]
4.7.0-rc5+ #524 Tainted: G           O
-------------------------------------------------------
kworker/u8:0/6 is trying to acquire lock:
 (&dev->mode_config.mutex){+.+.+.}, at: [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120

but task is already holding lock:
 ((fb_notifier_list).rwsem){++++.+}, at: [<ffffffff810ac195>] __blocking_notifier_call_chain+0x35/0x70

which lock already depends on the new lock.

the existing dependency chain (in reverse order) is:

-> #1 ((fb_notifier_list).rwsem){++++.+}:
       [<ffffffff810df611>] lock_acquire+0xb1/0x200
       [<ffffffff819a55b4>] down_write+0x44/0x80
       [<ffffffff810abf91>] blocking_notifier_chain_register+0x21/0xb0
       [<ffffffff814c7448>] fb_register_client+0x18/0x20
       [<ffffffff814c6c86>] backlight_device_register+0x136/0x260
       [<ffffffffa0127eb2>] intel_backlight_device_register+0xa2/0x160 [i915]
       [<ffffffffa00f46be>] intel_connector_register+0xe/0x10 [i915]
       [<ffffffffa0112bfb>] intel_dp_connector_register+0x1b/0x80 [i915]
       [<ffffffff8159dfea>] drm_connector_register+0x4a/0x80
       [<ffffffff8159fe44>] drm_connector_register_all+0x64/0xf0
       [<ffffffff815a2a64>] drm_modeset_register_all+0x174/0x1c0
       [<ffffffff81599b72>] drm_dev_register+0xc2/0xd0
       [<ffffffffa00621d7>] i915_driver_load+0x1547/0x2200 [i915]
       [<ffffffffa006d80f>] i915_pci_probe+0x4f/0x70 [i915]
       [<ffffffff814a2135>] local_pci_probe+0x45/0xa0
       [<ffffffff814a349b>] pci_device_probe+0xdb/0x130
       [<ffffffff815c07e3>] driver_probe_device+0x223/0x440
       [<ffffffff815c0ad5>] __driver_attach+0xd5/0x100
       [<ffffffff815be386>] bus_for_each_dev+0x66/0xa0
       [<ffffffff815c002e>] driver_attach+0x1e/0x20
       [<ffffffff815bf9be>] bus_add_driver+0x1ee/0x280
       [<ffffffff815c1810>] driver_register+0x60/0xe0
       [<ffffffff814a1a10>] __pci_register_driver+0x60/0x70
       [<ffffffffa01a905b>] i915_init+0x5b/0x62 [i915]
       [<ffffffff8100042d>] do_one_initcall+0x3d/0x150
       [<ffffffff811a935b>] do_init_module+0x5f/0x1d9
       [<ffffffff81124416>] load_module+0x20e6/0x27e0
       [<ffffffff81124d63>] SYSC_finit_module+0xc3/0xf0
       [<ffffffff81124dae>] SyS_finit_module+0xe/0x10
       [<ffffffff819a83a9>] entry_SYSCALL_64_fastpath+0x1c/0xac

-> #0 (&dev->mode_config.mutex){+.+.+.}:
       [<ffffffff810df0ac>] __lock_acquire+0x10fc/0x1260
       [<ffffffff810df611>] lock_acquire+0xb1/0x200
       [<ffffffff819a3097>] mutex_lock_nested+0x67/0x3c0
       [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120
       [<ffffffff8158f79b>] drm_fb_helper_restore_fbdev_mode_unlocked+0x2b/0x80
       [<ffffffff8158f81d>] drm_fb_helper_set_par+0x2d/0x50
       [<ffffffffa0105f7a>] intel_fbdev_set_par+0x1a/0x60 [i915]
       [<ffffffff814c13c6>] fbcon_init+0x586/0x610
       [<ffffffff8154d16a>] visual_init+0xca/0x130
       [<ffffffff8154e611>] do_bind_con_driver+0x1c1/0x3a0
       [<ffffffff8154eaf6>] do_take_over_console+0x116/0x180
       [<ffffffff814bd3a7>] do_fbcon_takeover+0x57/0xb0
       [<ffffffff814c1e48>] fbcon_event_notify+0x658/0x750
       [<ffffffff810abcae>] notifier_call_chain+0x3e/0xb0
       [<ffffffff810ac1ad>] __blocking_notifier_call_chain+0x4d/0x70
       [<ffffffff810ac1e6>] blocking_notifier_call_chain+0x16/0x20
       [<ffffffff814c748b>] fb_notifier_call_chain+0x1b/0x20
       [<ffffffff814c86b1>] register_framebuffer+0x251/0x330
       [<ffffffff8158fa9f>] drm_fb_helper_initial_config+0x25f/0x3f0
       [<ffffffffa0106b48>] intel_fbdev_initial_config+0x18/0x30 [i915]
       [<ffffffff810adfd8>] async_run_entry_fn+0x48/0x150
       [<ffffffff810a3947>] process_one_work+0x1e7/0x750
       [<ffffffff810a3efb>] worker_thread+0x4b/0x4f0
       [<ffffffff810aad4f>] kthread+0xef/0x110
       [<ffffffff819a85ef>] ret_from_fork+0x1f/0x40

other info that might help us debug this:

 Possible unsafe locking scenario:

       CPU0                    CPU1
       ----                    ----
  lock((fb_notifier_list).rwsem);
                               lock(&dev->mode_config.mutex);
                               lock((fb_notifier_list).rwsem);
  lock(&dev->mode_config.mutex);

 *** DEADLOCK ***

6 locks held by kworker/u8:0/6:
 #0:  ("events_unbound"){.+.+.+}, at: [<ffffffff810a38c9>] process_one_work+0x169/0x750
 #1:  ((&entry->work)){+.+.+.}, at: [<ffffffff810a38c9>] process_one_work+0x169/0x750
 #2:  (registration_lock){+.+.+.}, at: [<ffffffff814c8487>] register_framebuffer+0x27/0x330
 #3:  (console_lock){+.+.+.}, at: [<ffffffff814c86ce>] register_framebuffer+0x26e/0x330
 #4:  (&fb_info->lock){+.+.+.}, at: [<ffffffff814c78dd>] lock_fb_info+0x1d/0x40
 #5:  ((fb_notifier_list).rwsem){++++.+}, at: [<ffffffff810ac195>] __blocking_notifier_call_chain+0x35/0x70

stack backtrace:
CPU: 2 PID: 6 Comm: kworker/u8:0 Tainted: G           O    4.7.0-rc5+ #524
Hardware name: Intel Corp. Broxton P/NOTEBOOK, BIOS APLKRVPA.X64.0138.B33.1606250842 06/25/2016
Workqueue: events_unbound async_run_entry_fn
 0000000000000000 ffff8800758577f0 ffffffff814507a5 ffffffff828b9900
 ffffffff828b9900 ffff880075857830 ffffffff810dc6fa ffff880075857880
 ffff88007584d688 0000000000000005 0000000000000006 ffff88007584d6b0
Call Trace:
 [<ffffffff814507a5>] dump_stack+0x67/0x92
 [<ffffffff810dc6fa>] print_circular_bug+0x1aa/0x200
 [<ffffffff810df0ac>] __lock_acquire+0x10fc/0x1260
 [<ffffffff810df611>] lock_acquire+0xb1/0x200
 [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
 [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
 [<ffffffff819a3097>] mutex_lock_nested+0x67/0x3c0
 [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
 [<ffffffff810fa85f>] ? rcu_read_lock_sched_held+0x7f/0x90
 [<ffffffff81208218>] ? kmem_cache_alloc_trace+0x248/0x2b0
 [<ffffffff815afdc5>] ? drm_modeset_lock_all+0x25/0x120
 [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120
 [<ffffffff8158f79b>] drm_fb_helper_restore_fbdev_mode_unlocked+0x2b/0x80
 [<ffffffff8158f81d>] drm_fb_helper_set_par+0x2d/0x50
 [<ffffffffa0105f7a>] intel_fbdev_set_par+0x1a/0x60 [i915]
 [<ffffffff814c13c6>] fbcon_init+0x586/0x610
 [<ffffffff8154d16a>] visual_init+0xca/0x130
 [<ffffffff8154e611>] do_bind_con_driver+0x1c1/0x3a0
 [<ffffffff8154eaf6>] do_take_over_console+0x116/0x180
 [<ffffffff814bd3a7>] do_fbcon_takeover+0x57/0xb0
 [<ffffffff814c1e48>] fbcon_event_notify+0x658/0x750
 [<ffffffff810abcae>] notifier_call_chain+0x3e/0xb0
 [<ffffffff810ac1ad>] __blocking_notifier_call_chain+0x4d/0x70
 [<ffffffff810ac1e6>] blocking_notifier_call_chain+0x16/0x20
 [<ffffffff814c748b>] fb_notifier_call_chain+0x1b/0x20
 [<ffffffff814c86b1>] register_framebuffer+0x251/0x330
 [<ffffffff815b7e8d>] ? vga_switcheroo_client_fb_set+0x5d/0x70
 [<ffffffff8158fa9f>] drm_fb_helper_initial_config+0x25f/0x3f0
 [<ffffffffa0106b48>] intel_fbdev_initial_config+0x18/0x30 [i915]
 [<ffffffff810adfd8>] async_run_entry_fn+0x48/0x150
 [<ffffffff810a3947>] process_one_work+0x1e7/0x750
 [<ffffffff810a38c9>] ? process_one_work+0x169/0x750
 [<ffffffff810a3efb>] worker_thread+0x4b/0x4f0
 [<ffffffff810a3eb0>] ? process_one_work+0x750/0x750
 [<ffffffff810aad4f>] kthread+0xef/0x110
 [<ffffffff819a85ef>] ret_from_fork+0x1f/0x40
 [<ffffffff810aac60>] ? kthread_stop+0x2e0/0x2e0

Cc: Imre Deak <imre.deak@intel.com>
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
---
 drivers/gpu/drm/drm_connector.c | 8 +++-----
 1 file changed, 3 insertions(+), 5 deletions(-)

diff --git a/drivers/gpu/drm/drm_connector.c b/drivers/gpu/drm/drm_connector.c
index ef921fa09a84..d9104d8b3c6b 100644
--- a/drivers/gpu/drm/drm_connector.c
+++ b/drivers/gpu/drm/drm_connector.c
@@ -401,16 +401,14 @@ int drm_connector_register_all(struct drm_device *dev)
 	struct drm_connector *connector;
 	int ret;
 
-	mutex_lock(&dev->mode_config.mutex);
-
-	drm_for_each_connector(connector, dev) {
+	/* FIXME: taking the mode config mutex ends up in a clash with
+	 * fbcon/backlight registration */
+	list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
 		ret = drm_connector_register(connector);
 		if (ret)
 			goto err;
 	}
 
-	mutex_unlock(&dev->mode_config.mutex);
-
 	return 0;
 
 err:
-- 
2.8.1

_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel

^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [PATCH] drm: Paper over locking inversion after registration rework
  2016-08-04 10:15 [PATCH] drm: Paper over locking inversion after registration rework Daniel Vetter
@ 2016-08-04 10:35 ` Jani Nikula
  2016-08-04 11:12 ` ✗ Ro.CI.BAT: failure for " Patchwork
  2016-08-04 16:13 ` [PATCH] " Chris Wilson
  2 siblings, 0 replies; 7+ messages in thread
From: Jani Nikula @ 2016-08-04 10:35 UTC (permalink / raw)
  To: DRI Development; +Cc: Daniel Vetter, Intel Graphics Development, Daniel Vetter

On Thu, 04 Aug 2016, Daniel Vetter <daniel.vetter@ffwll.ch> wrote:
> drm_connector_register_all requires a few too many locks because our
> connector_list locking is busted. Add another FIXME+hack to work
> around this. This should address the below lockdep splat:
>
> ======================================================
> [ INFO: possible circular locking dependency detected ]
> 4.7.0-rc5+ #524 Tainted: G           O
> -------------------------------------------------------
> kworker/u8:0/6 is trying to acquire lock:
>  (&dev->mode_config.mutex){+.+.+.}, at: [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120
>
> but task is already holding lock:
>  ((fb_notifier_list).rwsem){++++.+}, at: [<ffffffff810ac195>] __blocking_notifier_call_chain+0x35/0x70
>
> which lock already depends on the new lock.
>
> the existing dependency chain (in reverse order) is:
>
> -> #1 ((fb_notifier_list).rwsem){++++.+}:
>        [<ffffffff810df611>] lock_acquire+0xb1/0x200
>        [<ffffffff819a55b4>] down_write+0x44/0x80
>        [<ffffffff810abf91>] blocking_notifier_chain_register+0x21/0xb0
>        [<ffffffff814c7448>] fb_register_client+0x18/0x20
>        [<ffffffff814c6c86>] backlight_device_register+0x136/0x260
>        [<ffffffffa0127eb2>] intel_backlight_device_register+0xa2/0x160 [i915]
>        [<ffffffffa00f46be>] intel_connector_register+0xe/0x10 [i915]
>        [<ffffffffa0112bfb>] intel_dp_connector_register+0x1b/0x80 [i915]
>        [<ffffffff8159dfea>] drm_connector_register+0x4a/0x80
>        [<ffffffff8159fe44>] drm_connector_register_all+0x64/0xf0
>        [<ffffffff815a2a64>] drm_modeset_register_all+0x174/0x1c0
>        [<ffffffff81599b72>] drm_dev_register+0xc2/0xd0
>        [<ffffffffa00621d7>] i915_driver_load+0x1547/0x2200 [i915]
>        [<ffffffffa006d80f>] i915_pci_probe+0x4f/0x70 [i915]
>        [<ffffffff814a2135>] local_pci_probe+0x45/0xa0
>        [<ffffffff814a349b>] pci_device_probe+0xdb/0x130
>        [<ffffffff815c07e3>] driver_probe_device+0x223/0x440
>        [<ffffffff815c0ad5>] __driver_attach+0xd5/0x100
>        [<ffffffff815be386>] bus_for_each_dev+0x66/0xa0
>        [<ffffffff815c002e>] driver_attach+0x1e/0x20
>        [<ffffffff815bf9be>] bus_add_driver+0x1ee/0x280
>        [<ffffffff815c1810>] driver_register+0x60/0xe0
>        [<ffffffff814a1a10>] __pci_register_driver+0x60/0x70
>        [<ffffffffa01a905b>] i915_init+0x5b/0x62 [i915]
>        [<ffffffff8100042d>] do_one_initcall+0x3d/0x150
>        [<ffffffff811a935b>] do_init_module+0x5f/0x1d9
>        [<ffffffff81124416>] load_module+0x20e6/0x27e0
>        [<ffffffff81124d63>] SYSC_finit_module+0xc3/0xf0
>        [<ffffffff81124dae>] SyS_finit_module+0xe/0x10
>        [<ffffffff819a83a9>] entry_SYSCALL_64_fastpath+0x1c/0xac
>
> -> #0 (&dev->mode_config.mutex){+.+.+.}:
>        [<ffffffff810df0ac>] __lock_acquire+0x10fc/0x1260
>        [<ffffffff810df611>] lock_acquire+0xb1/0x200
>        [<ffffffff819a3097>] mutex_lock_nested+0x67/0x3c0
>        [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120
>        [<ffffffff8158f79b>] drm_fb_helper_restore_fbdev_mode_unlocked+0x2b/0x80
>        [<ffffffff8158f81d>] drm_fb_helper_set_par+0x2d/0x50
>        [<ffffffffa0105f7a>] intel_fbdev_set_par+0x1a/0x60 [i915]
>        [<ffffffff814c13c6>] fbcon_init+0x586/0x610
>        [<ffffffff8154d16a>] visual_init+0xca/0x130
>        [<ffffffff8154e611>] do_bind_con_driver+0x1c1/0x3a0
>        [<ffffffff8154eaf6>] do_take_over_console+0x116/0x180
>        [<ffffffff814bd3a7>] do_fbcon_takeover+0x57/0xb0
>        [<ffffffff814c1e48>] fbcon_event_notify+0x658/0x750
>        [<ffffffff810abcae>] notifier_call_chain+0x3e/0xb0
>        [<ffffffff810ac1ad>] __blocking_notifier_call_chain+0x4d/0x70
>        [<ffffffff810ac1e6>] blocking_notifier_call_chain+0x16/0x20
>        [<ffffffff814c748b>] fb_notifier_call_chain+0x1b/0x20
>        [<ffffffff814c86b1>] register_framebuffer+0x251/0x330
>        [<ffffffff8158fa9f>] drm_fb_helper_initial_config+0x25f/0x3f0
>        [<ffffffffa0106b48>] intel_fbdev_initial_config+0x18/0x30 [i915]
>        [<ffffffff810adfd8>] async_run_entry_fn+0x48/0x150
>        [<ffffffff810a3947>] process_one_work+0x1e7/0x750
>        [<ffffffff810a3efb>] worker_thread+0x4b/0x4f0
>        [<ffffffff810aad4f>] kthread+0xef/0x110
>        [<ffffffff819a85ef>] ret_from_fork+0x1f/0x40
>
> other info that might help us debug this:
>
>  Possible unsafe locking scenario:
>
>        CPU0                    CPU1
>        ----                    ----
>   lock((fb_notifier_list).rwsem);
>                                lock(&dev->mode_config.mutex);
>                                lock((fb_notifier_list).rwsem);
>   lock(&dev->mode_config.mutex);
>
>  *** DEADLOCK ***
>
> 6 locks held by kworker/u8:0/6:
>  #0:  ("events_unbound"){.+.+.+}, at: [<ffffffff810a38c9>] process_one_work+0x169/0x750
>  #1:  ((&entry->work)){+.+.+.}, at: [<ffffffff810a38c9>] process_one_work+0x169/0x750
>  #2:  (registration_lock){+.+.+.}, at: [<ffffffff814c8487>] register_framebuffer+0x27/0x330
>  #3:  (console_lock){+.+.+.}, at: [<ffffffff814c86ce>] register_framebuffer+0x26e/0x330
>  #4:  (&fb_info->lock){+.+.+.}, at: [<ffffffff814c78dd>] lock_fb_info+0x1d/0x40
>  #5:  ((fb_notifier_list).rwsem){++++.+}, at: [<ffffffff810ac195>] __blocking_notifier_call_chain+0x35/0x70
>
> stack backtrace:
> CPU: 2 PID: 6 Comm: kworker/u8:0 Tainted: G           O    4.7.0-rc5+ #524
> Hardware name: Intel Corp. Broxton P/NOTEBOOK, BIOS APLKRVPA.X64.0138.B33.1606250842 06/25/2016
> Workqueue: events_unbound async_run_entry_fn
>  0000000000000000 ffff8800758577f0 ffffffff814507a5 ffffffff828b9900
>  ffffffff828b9900 ffff880075857830 ffffffff810dc6fa ffff880075857880
>  ffff88007584d688 0000000000000005 0000000000000006 ffff88007584d6b0
> Call Trace:
>  [<ffffffff814507a5>] dump_stack+0x67/0x92
>  [<ffffffff810dc6fa>] print_circular_bug+0x1aa/0x200
>  [<ffffffff810df0ac>] __lock_acquire+0x10fc/0x1260
>  [<ffffffff810df611>] lock_acquire+0xb1/0x200
>  [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
>  [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
>  [<ffffffff819a3097>] mutex_lock_nested+0x67/0x3c0
>  [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
>  [<ffffffff810fa85f>] ? rcu_read_lock_sched_held+0x7f/0x90
>  [<ffffffff81208218>] ? kmem_cache_alloc_trace+0x248/0x2b0
>  [<ffffffff815afdc5>] ? drm_modeset_lock_all+0x25/0x120
>  [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120
>  [<ffffffff8158f79b>] drm_fb_helper_restore_fbdev_mode_unlocked+0x2b/0x80
>  [<ffffffff8158f81d>] drm_fb_helper_set_par+0x2d/0x50
>  [<ffffffffa0105f7a>] intel_fbdev_set_par+0x1a/0x60 [i915]
>  [<ffffffff814c13c6>] fbcon_init+0x586/0x610
>  [<ffffffff8154d16a>] visual_init+0xca/0x130
>  [<ffffffff8154e611>] do_bind_con_driver+0x1c1/0x3a0
>  [<ffffffff8154eaf6>] do_take_over_console+0x116/0x180
>  [<ffffffff814bd3a7>] do_fbcon_takeover+0x57/0xb0
>  [<ffffffff814c1e48>] fbcon_event_notify+0x658/0x750
>  [<ffffffff810abcae>] notifier_call_chain+0x3e/0xb0
>  [<ffffffff810ac1ad>] __blocking_notifier_call_chain+0x4d/0x70
>  [<ffffffff810ac1e6>] blocking_notifier_call_chain+0x16/0x20
>  [<ffffffff814c748b>] fb_notifier_call_chain+0x1b/0x20
>  [<ffffffff814c86b1>] register_framebuffer+0x251/0x330
>  [<ffffffff815b7e8d>] ? vga_switcheroo_client_fb_set+0x5d/0x70
>  [<ffffffff8158fa9f>] drm_fb_helper_initial_config+0x25f/0x3f0
>  [<ffffffffa0106b48>] intel_fbdev_initial_config+0x18/0x30 [i915]
>  [<ffffffff810adfd8>] async_run_entry_fn+0x48/0x150
>  [<ffffffff810a3947>] process_one_work+0x1e7/0x750
>  [<ffffffff810a38c9>] ? process_one_work+0x169/0x750
>  [<ffffffff810a3efb>] worker_thread+0x4b/0x4f0
>  [<ffffffff810a3eb0>] ? process_one_work+0x750/0x750
>  [<ffffffff810aad4f>] kthread+0xef/0x110
>  [<ffffffff819a85ef>] ret_from_fork+0x1f/0x40
>  [<ffffffff810aac60>] ? kthread_stop+0x2e0/0x2e0
>
> Cc: Imre Deak <imre.deak@intel.com>
> Cc: Chris Wilson <chris@chris-wilson.co.uk>
> Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>

Fixes: ?
Cc: drm-intel-fixes@lists.freedesktop.org
Cc: a bunch of people who reported the problem upstream?

BR,
Jani.

> ---
>  drivers/gpu/drm/drm_connector.c | 8 +++-----
>  1 file changed, 3 insertions(+), 5 deletions(-)
>
> diff --git a/drivers/gpu/drm/drm_connector.c b/drivers/gpu/drm/drm_connector.c
> index ef921fa09a84..d9104d8b3c6b 100644
> --- a/drivers/gpu/drm/drm_connector.c
> +++ b/drivers/gpu/drm/drm_connector.c
> @@ -401,16 +401,14 @@ int drm_connector_register_all(struct drm_device *dev)
>  	struct drm_connector *connector;
>  	int ret;
>  
> -	mutex_lock(&dev->mode_config.mutex);
> -
> -	drm_for_each_connector(connector, dev) {
> +	/* FIXME: taking the mode config mutex ends up in a clash with
> +	 * fbcon/backlight registration */
> +	list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
>  		ret = drm_connector_register(connector);
>  		if (ret)
>  			goto err;
>  	}
>  
> -	mutex_unlock(&dev->mode_config.mutex);
> -
>  	return 0;
>  
>  err:

-- 
Jani Nikula, Intel Open Source Technology Center
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

^ permalink raw reply	[flat|nested] 7+ messages in thread

* ✗ Ro.CI.BAT: failure for drm: Paper over locking inversion after registration rework
  2016-08-04 10:15 [PATCH] drm: Paper over locking inversion after registration rework Daniel Vetter
  2016-08-04 10:35 ` Jani Nikula
@ 2016-08-04 11:12 ` Patchwork
  2016-08-04 16:13 ` [PATCH] " Chris Wilson
  2 siblings, 0 replies; 7+ messages in thread
From: Patchwork @ 2016-08-04 11:12 UTC (permalink / raw)
  To: Daniel Vetter; +Cc: intel-gfx

== Series Details ==

Series: drm: Paper over locking inversion after registration rework
URL   : https://patchwork.freedesktop.org/series/10653/
State : failure

== Summary ==

Applying: drm: Paper over locking inversion after registration rework
fatal: sha1 information is lacking or useless (drivers/gpu/drm/drm_connector.c).
error: could not build fake ancestor
Patch failed at 0001 drm: Paper over locking inversion after registration rework
The copy of the patch that failed is found in: .git/rebase-apply/patch
When you have resolved this problem, run "git am --continue".
If you prefer to skip this patch, run "git am --skip" instead.
To restore the original branch and stop patching, run "git am --abort".

_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH] drm: Paper over locking inversion after registration rework
  2016-08-04 10:15 [PATCH] drm: Paper over locking inversion after registration rework Daniel Vetter
  2016-08-04 10:35 ` Jani Nikula
  2016-08-04 11:12 ` ✗ Ro.CI.BAT: failure for " Patchwork
@ 2016-08-04 16:13 ` Chris Wilson
  2 siblings, 0 replies; 7+ messages in thread
From: Chris Wilson @ 2016-08-04 16:13 UTC (permalink / raw)
  To: Daniel Vetter; +Cc: Daniel Vetter, Intel Graphics Development, DRI Development

On Thu, Aug 04, 2016 at 12:15:14PM +0200, Daniel Vetter wrote:
> drm_connector_register_all requires a few too many locks because our
> connector_list locking is busted. Add another FIXME+hack to work
> around this. This should address the below lockdep splat:
> 
> Cc: Imre Deak <imre.deak@intel.com>
> Cc: Chris Wilson <chris@chris-wilson.co.uk>
> Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
> ---
>  drivers/gpu/drm/drm_connector.c | 8 +++-----
>  1 file changed, 3 insertions(+), 5 deletions(-)
> 
> diff --git a/drivers/gpu/drm/drm_connector.c b/drivers/gpu/drm/drm_connector.c
> index ef921fa09a84..d9104d8b3c6b 100644
> --- a/drivers/gpu/drm/drm_connector.c
> +++ b/drivers/gpu/drm/drm_connector.c
> @@ -401,16 +401,14 @@ int drm_connector_register_all(struct drm_device *dev)
>  	struct drm_connector *connector;
>  	int ret;
>  
> -	mutex_lock(&dev->mode_config.mutex);
> -
> -	drm_for_each_connector(connector, dev) {
> +	/* FIXME: taking the mode config mutex ends up in a clash with
> +	 * fbcon/backlight registration */
> +	list_for_each_entry(connector, &dev->mode_config.connector_list, head) {

Ok, double checked that the only time we delete from this list
(currently) is in cleanup. That's highly unlikely to be running at the
same time as the register, so

Acked-by: Chris Wilson <chris@chris-wilson.co.uk>
-Chris

-- 
Chris Wilson, Intel Open Source Technology Centre
_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel

^ permalink raw reply	[flat|nested] 7+ messages in thread

* [PATCH] drm: Paper over locking inversion after registration rework
@ 2016-08-04 16:35 Daniel Vetter
  2016-08-05  8:57 ` Jiri Kosina
  0 siblings, 1 reply; 7+ messages in thread
From: Daniel Vetter @ 2016-08-04 16:35 UTC (permalink / raw)
  To: DRI Development
  Cc: Daniel Vetter, Intel Graphics Development, Jiri Kosina,
	Daniel Vetter

drm_connector_register_all requires a few too many locks because our
connector_list locking is busted. Add another FIXME+hack to work
around this. This should address the below lockdep splat:

======================================================
[ INFO: possible circular locking dependency detected ]
4.7.0-rc5+ #524 Tainted: G           O
-------------------------------------------------------
kworker/u8:0/6 is trying to acquire lock:
 (&dev->mode_config.mutex){+.+.+.}, at: [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120

but task is already holding lock:
 ((fb_notifier_list).rwsem){++++.+}, at: [<ffffffff810ac195>] __blocking_notifier_call_chain+0x35/0x70

which lock already depends on the new lock.

the existing dependency chain (in reverse order) is:

-> #1 ((fb_notifier_list).rwsem){++++.+}:
       [<ffffffff810df611>] lock_acquire+0xb1/0x200
       [<ffffffff819a55b4>] down_write+0x44/0x80
       [<ffffffff810abf91>] blocking_notifier_chain_register+0x21/0xb0
       [<ffffffff814c7448>] fb_register_client+0x18/0x20
       [<ffffffff814c6c86>] backlight_device_register+0x136/0x260
       [<ffffffffa0127eb2>] intel_backlight_device_register+0xa2/0x160 [i915]
       [<ffffffffa00f46be>] intel_connector_register+0xe/0x10 [i915]
       [<ffffffffa0112bfb>] intel_dp_connector_register+0x1b/0x80 [i915]
       [<ffffffff8159dfea>] drm_connector_register+0x4a/0x80
       [<ffffffff8159fe44>] drm_connector_register_all+0x64/0xf0
       [<ffffffff815a2a64>] drm_modeset_register_all+0x174/0x1c0
       [<ffffffff81599b72>] drm_dev_register+0xc2/0xd0
       [<ffffffffa00621d7>] i915_driver_load+0x1547/0x2200 [i915]
       [<ffffffffa006d80f>] i915_pci_probe+0x4f/0x70 [i915]
       [<ffffffff814a2135>] local_pci_probe+0x45/0xa0
       [<ffffffff814a349b>] pci_device_probe+0xdb/0x130
       [<ffffffff815c07e3>] driver_probe_device+0x223/0x440
       [<ffffffff815c0ad5>] __driver_attach+0xd5/0x100
       [<ffffffff815be386>] bus_for_each_dev+0x66/0xa0
       [<ffffffff815c002e>] driver_attach+0x1e/0x20
       [<ffffffff815bf9be>] bus_add_driver+0x1ee/0x280
       [<ffffffff815c1810>] driver_register+0x60/0xe0
       [<ffffffff814a1a10>] __pci_register_driver+0x60/0x70
       [<ffffffffa01a905b>] i915_init+0x5b/0x62 [i915]
       [<ffffffff8100042d>] do_one_initcall+0x3d/0x150
       [<ffffffff811a935b>] do_init_module+0x5f/0x1d9
       [<ffffffff81124416>] load_module+0x20e6/0x27e0
       [<ffffffff81124d63>] SYSC_finit_module+0xc3/0xf0
       [<ffffffff81124dae>] SyS_finit_module+0xe/0x10
       [<ffffffff819a83a9>] entry_SYSCALL_64_fastpath+0x1c/0xac

-> #0 (&dev->mode_config.mutex){+.+.+.}:
       [<ffffffff810df0ac>] __lock_acquire+0x10fc/0x1260
       [<ffffffff810df611>] lock_acquire+0xb1/0x200
       [<ffffffff819a3097>] mutex_lock_nested+0x67/0x3c0
       [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120
       [<ffffffff8158f79b>] drm_fb_helper_restore_fbdev_mode_unlocked+0x2b/0x80
       [<ffffffff8158f81d>] drm_fb_helper_set_par+0x2d/0x50
       [<ffffffffa0105f7a>] intel_fbdev_set_par+0x1a/0x60 [i915]
       [<ffffffff814c13c6>] fbcon_init+0x586/0x610
       [<ffffffff8154d16a>] visual_init+0xca/0x130
       [<ffffffff8154e611>] do_bind_con_driver+0x1c1/0x3a0
       [<ffffffff8154eaf6>] do_take_over_console+0x116/0x180
       [<ffffffff814bd3a7>] do_fbcon_takeover+0x57/0xb0
       [<ffffffff814c1e48>] fbcon_event_notify+0x658/0x750
       [<ffffffff810abcae>] notifier_call_chain+0x3e/0xb0
       [<ffffffff810ac1ad>] __blocking_notifier_call_chain+0x4d/0x70
       [<ffffffff810ac1e6>] blocking_notifier_call_chain+0x16/0x20
       [<ffffffff814c748b>] fb_notifier_call_chain+0x1b/0x20
       [<ffffffff814c86b1>] register_framebuffer+0x251/0x330
       [<ffffffff8158fa9f>] drm_fb_helper_initial_config+0x25f/0x3f0
       [<ffffffffa0106b48>] intel_fbdev_initial_config+0x18/0x30 [i915]
       [<ffffffff810adfd8>] async_run_entry_fn+0x48/0x150
       [<ffffffff810a3947>] process_one_work+0x1e7/0x750
       [<ffffffff810a3efb>] worker_thread+0x4b/0x4f0
       [<ffffffff810aad4f>] kthread+0xef/0x110
       [<ffffffff819a85ef>] ret_from_fork+0x1f/0x40

other info that might help us debug this:

 Possible unsafe locking scenario:

       CPU0                    CPU1
       ----                    ----
  lock((fb_notifier_list).rwsem);
                               lock(&dev->mode_config.mutex);
                               lock((fb_notifier_list).rwsem);
  lock(&dev->mode_config.mutex);

 *** DEADLOCK ***

6 locks held by kworker/u8:0/6:
 #0:  ("events_unbound"){.+.+.+}, at: [<ffffffff810a38c9>] process_one_work+0x169/0x750
 #1:  ((&entry->work)){+.+.+.}, at: [<ffffffff810a38c9>] process_one_work+0x169/0x750
 #2:  (registration_lock){+.+.+.}, at: [<ffffffff814c8487>] register_framebuffer+0x27/0x330
 #3:  (console_lock){+.+.+.}, at: [<ffffffff814c86ce>] register_framebuffer+0x26e/0x330
 #4:  (&fb_info->lock){+.+.+.}, at: [<ffffffff814c78dd>] lock_fb_info+0x1d/0x40
 #5:  ((fb_notifier_list).rwsem){++++.+}, at: [<ffffffff810ac195>] __blocking_notifier_call_chain+0x35/0x70

stack backtrace:
CPU: 2 PID: 6 Comm: kworker/u8:0 Tainted: G           O    4.7.0-rc5+ #524
Hardware name: Intel Corp. Broxton P/NOTEBOOK, BIOS APLKRVPA.X64.0138.B33.1606250842 06/25/2016
Workqueue: events_unbound async_run_entry_fn
 0000000000000000 ffff8800758577f0 ffffffff814507a5 ffffffff828b9900
 ffffffff828b9900 ffff880075857830 ffffffff810dc6fa ffff880075857880
 ffff88007584d688 0000000000000005 0000000000000006 ffff88007584d6b0
Call Trace:
 [<ffffffff814507a5>] dump_stack+0x67/0x92
 [<ffffffff810dc6fa>] print_circular_bug+0x1aa/0x200
 [<ffffffff810df0ac>] __lock_acquire+0x10fc/0x1260
 [<ffffffff810df611>] lock_acquire+0xb1/0x200
 [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
 [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
 [<ffffffff819a3097>] mutex_lock_nested+0x67/0x3c0
 [<ffffffff815afde0>] ? drm_modeset_lock_all+0x40/0x120
 [<ffffffff810fa85f>] ? rcu_read_lock_sched_held+0x7f/0x90
 [<ffffffff81208218>] ? kmem_cache_alloc_trace+0x248/0x2b0
 [<ffffffff815afdc5>] ? drm_modeset_lock_all+0x25/0x120
 [<ffffffff815afde0>] drm_modeset_lock_all+0x40/0x120
 [<ffffffff8158f79b>] drm_fb_helper_restore_fbdev_mode_unlocked+0x2b/0x80
 [<ffffffff8158f81d>] drm_fb_helper_set_par+0x2d/0x50
 [<ffffffffa0105f7a>] intel_fbdev_set_par+0x1a/0x60 [i915]
 [<ffffffff814c13c6>] fbcon_init+0x586/0x610
 [<ffffffff8154d16a>] visual_init+0xca/0x130
 [<ffffffff8154e611>] do_bind_con_driver+0x1c1/0x3a0
 [<ffffffff8154eaf6>] do_take_over_console+0x116/0x180
 [<ffffffff814bd3a7>] do_fbcon_takeover+0x57/0xb0
 [<ffffffff814c1e48>] fbcon_event_notify+0x658/0x750
 [<ffffffff810abcae>] notifier_call_chain+0x3e/0xb0
 [<ffffffff810ac1ad>] __blocking_notifier_call_chain+0x4d/0x70
 [<ffffffff810ac1e6>] blocking_notifier_call_chain+0x16/0x20
 [<ffffffff814c748b>] fb_notifier_call_chain+0x1b/0x20
 [<ffffffff814c86b1>] register_framebuffer+0x251/0x330
 [<ffffffff815b7e8d>] ? vga_switcheroo_client_fb_set+0x5d/0x70
 [<ffffffff8158fa9f>] drm_fb_helper_initial_config+0x25f/0x3f0
 [<ffffffffa0106b48>] intel_fbdev_initial_config+0x18/0x30 [i915]
 [<ffffffff810adfd8>] async_run_entry_fn+0x48/0x150
 [<ffffffff810a3947>] process_one_work+0x1e7/0x750
 [<ffffffff810a38c9>] ? process_one_work+0x169/0x750
 [<ffffffff810a3efb>] worker_thread+0x4b/0x4f0
 [<ffffffff810a3eb0>] ? process_one_work+0x750/0x750
 [<ffffffff810aad4f>] kthread+0xef/0x110
 [<ffffffff819a85ef>] ret_from_fork+0x1f/0x40
 [<ffffffff810aac60>] ? kthread_stop+0x2e0/0x2e0

v2: Rebase onto the right branch (hand-editing patches ftw) and add more
reporters.

Reported-by: Imre Deak <imre.deak@intel.com>
Cc: Imre Deak <imre.deak@intel.com>
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Acked-by: Chris Wilson <chris@chris-wilson.co.uk>
Reported-by: Jiri Kosina <jikos@kernel.org>
Cc: Jiri Kosina <jikos@kernel.org>
Signed-off-by: Daniel Vetter <daniel.vetter@intel.com>
---
 drivers/gpu/drm/drm_crtc.c | 8 +++-----
 1 file changed, 3 insertions(+), 5 deletions(-)

diff --git a/drivers/gpu/drm/drm_crtc.c b/drivers/gpu/drm/drm_crtc.c
index ef921fa09a84..d9104d8b3c6b 100644
--- a/drivers/gpu/drm/drm_crtc.c
+++ b/drivers/gpu/drm/drm_crtc.c
@@ -401,16 +401,14 @@ int drm_connector_register_all(struct drm_device *dev)
 	struct drm_connector *connector;
 	int ret;
 
-	mutex_lock(&dev->mode_config.mutex);
-
-	drm_for_each_connector(connector, dev) {
+	/* FIXME: taking the mode config mutex ends up in a clash with
+	 * fbcon/backlight registration */
+	list_for_each_entry(connector, &dev->mode_config.connector_list, head) {
 		ret = drm_connector_register(connector);
 		if (ret)
 			goto err;
 	}
 
-	mutex_unlock(&dev->mode_config.mutex);
-
 	return 0;
 
 err:
-- 
2.8.1

_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

^ permalink raw reply related	[flat|nested] 7+ messages in thread

* Re: [PATCH] drm: Paper over locking inversion after registration rework
  2016-08-04 16:35 Daniel Vetter
@ 2016-08-05  8:57 ` Jiri Kosina
  2016-08-05  9:21   ` Daniel Vetter
  0 siblings, 1 reply; 7+ messages in thread
From: Jiri Kosina @ 2016-08-05  8:57 UTC (permalink / raw)
  To: Daniel Vetter; +Cc: Daniel Vetter, Intel Graphics Development, DRI Development

On Thu, 4 Aug 2016, Daniel Vetter wrote:

> drm_connector_register_all requires a few too many locks because our
> connector_list locking is busted. Add another FIXME+hack to work
> around this. This should address the below lockdep splat:
[ ... snip ... ]
> v2: Rebase onto the right branch (hand-editing patches ftw) and add more
> reporters.
> 
> Reported-by: Imre Deak <imre.deak@intel.com>
> Cc: Imre Deak <imre.deak@intel.com>
> Cc: Chris Wilson <chris@chris-wilson.co.uk>
> Acked-by: Chris Wilson <chris@chris-wilson.co.uk>
> Reported-by: Jiri Kosina <jikos@kernel.org>

Although I have no idea why this is safe thing to do :) I can confirm that 
it removes the lockdep complaint on my system. So in that regard,

	Tested-by: Jiri Kosina <jkosina@suse.cz>

Thanks,

-- 
Jiri Kosina
SUSE Labs

_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCH] drm: Paper over locking inversion after registration rework
  2016-08-05  8:57 ` Jiri Kosina
@ 2016-08-05  9:21   ` Daniel Vetter
  0 siblings, 0 replies; 7+ messages in thread
From: Daniel Vetter @ 2016-08-05  9:21 UTC (permalink / raw)
  To: Jiri Kosina, Dave Airlie
  Cc: Daniel Vetter, Intel Graphics Development, DRI Development,
	Daniel Vetter

On Fri, Aug 05, 2016 at 10:57:58AM +0200, Jiri Kosina wrote:
> On Thu, 4 Aug 2016, Daniel Vetter wrote:
> 
> > drm_connector_register_all requires a few too many locks because our
> > connector_list locking is busted. Add another FIXME+hack to work
> > around this. This should address the below lockdep splat:
> [ ... snip ... ]
> > v2: Rebase onto the right branch (hand-editing patches ftw) and add more
> > reporters.
> > 
> > Reported-by: Imre Deak <imre.deak@intel.com>
> > Cc: Imre Deak <imre.deak@intel.com>
> > Cc: Chris Wilson <chris@chris-wilson.co.uk>
> > Acked-by: Chris Wilson <chris@chris-wilson.co.uk>
> > Reported-by: Jiri Kosina <jikos@kernel.org>
> 
> Although I have no idea why this is safe thing to do :) I can confirm that 
> it removes the lockdep complaint on my system. So in that regard,
> 
> 	Tested-by: Jiri Kosina <jkosina@suse.cz>

Dave, can you pls pick this up for 4.8?
-Daniel
-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2016-08-05  9:22 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2016-08-04 10:15 [PATCH] drm: Paper over locking inversion after registration rework Daniel Vetter
2016-08-04 10:35 ` Jani Nikula
2016-08-04 11:12 ` ✗ Ro.CI.BAT: failure for " Patchwork
2016-08-04 16:13 ` [PATCH] " Chris Wilson
  -- strict thread matches above, loose matches on Subject: below --
2016-08-04 16:35 Daniel Vetter
2016-08-05  8:57 ` Jiri Kosina
2016-08-05  9:21   ` Daniel Vetter

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox