/* SPDX-License-Identifier: MIT */ /* * Copyright © 2019 Intel Corporation */ #ifndef __INTEL_GT_TYPES__ #define __INTEL_GT_TYPES__ #include #include #include #include #include #include #include "uc/intel_uc.h" #include "i915_vma.h" #include "intel_engine_types.h" #include "intel_reset_types.h" #include "intel_wakeref.h" struct drm_i915_private; struct i915_ggtt; struct intel_engine_cs; struct intel_uncore; struct intel_hangcheck { /* For hangcheck timer */ #define DRM_I915_HANGCHECK_PERIOD 1500 /* in ms */ #define DRM_I915_HANGCHECK_JIFFIES msecs_to_jiffies(DRM_I915_HANGCHECK_PERIOD) struct delayed_work work; }; struct intel_gt { struct drm_i915_private *i915; struct intel_uncore *uncore; struct i915_ggtt *ggtt; struct intel_uc uc; struct intel_gt_timelines { spinlock_t lock; /* protects active_list */ struct list_head active_list; /* Pack multiple timelines' seqnos into the same page */ spinlock_t hwsp_lock; struct list_head hwsp_free_list; } timelines; struct intel_wakeref wakeref; struct list_head closed_vma; spinlock_t closed_lock; /* guards the list of closed_vma */ struct intel_hangcheck hangcheck; struct intel_reset reset; /** * Is the GPU currently considered idle, or busy executing * userspace requests? Whilst idle, we allow runtime power * management to power down the hardware and display clocks. * In order to reduce the effect on performance, there * is a slight delay before we do so. */ intel_wakeref_t awake; struct blocking_notifier_head pm_notifications; ktime_t last_init_time; struct i915_vma *scratch; spinlock_t irq_lock; u32 gt_imr; u32 pm_ier; u32 pm_imr; u32 pm_guc_events; struct intel_engine_cs *engine[I915_NUM_ENGINES]; struct intel_engine_cs *engine_class[MAX_ENGINE_CLASS + 1] [MAX_ENGINE_INSTANCE + 1]; }; enum intel_gt_scratch_field { /* 8 bytes */ INTEL_GT_SCRATCH_FIELD_DEFAULT = 0, /* 8 bytes */ INTEL_GT_SCRATCH_FIELD_CLEAR_SLM_WA = 128, /* 8 bytes */ INTEL_GT_SCRATCH_FIELD_RENDER_FLUSH = 128, /* 8 bytes */ INTEL_GT_SCRATCH_FIELD_COHERENTL3_WA = 256, }; #endif /* __INTEL_GT_TYPES_H__ */