/* SPDX-License-Identifier: MIT */ /* * Copyright © 2019 Intel Corporation */ #ifndef __INTEL_GT_TYPES__ #define __INTEL_GT_TYPES__ #include #include #include #include #include #include #include #include #include "uc/intel_uc.h" #include "i915_vma.h" #include "intel_engine_types.h" #include "intel_gt_buffer_pool_types.h" #include "intel_llc_types.h" #include "intel_reset_types.h" #include "intel_rc6_types.h" #include "intel_rps_types.h" #include "intel_migrate_types.h" #include "intel_wakeref.h" #include "pxp/intel_pxp_types.h" struct drm_i915_private; struct i915_ggtt; struct intel_engine_cs; struct intel_uncore; struct intel_mmio_range { u32 start; u32 end; }; /* * The hardware has multiple kinds of multicast register ranges that need * special register steering (and future platforms are expected to add * additional types). * * During driver startup, we initialize the steering control register to * direct reads to a slice/subslice that are valid for the 'subslice' class * of multicast registers. If another type of steering does not have any * overlap in valid steering targets with 'subslice' style registers, we will * need to explicitly re-steer reads of registers of the other type. * * Only the replication types that may need additional non-default steering * are listed here. */ enum intel_steering_type { L3BANK, MSLICE, LNCF, NUM_STEERING_TYPES }; enum intel_submission_method { INTEL_SUBMISSION_RING, INTEL_SUBMISSION_ELSP, INTEL_SUBMISSION_GUC, }; struct intel_gt { struct drm_i915_private *i915; struct intel_uncore *uncore; struct i915_ggtt *ggtt; struct intel_uc uc; struct i915_wa_list wa_list; struct intel_gt_timelines { spinlock_t lock; /* protects active_list */ struct list_head active_list; } timelines; struct intel_gt_requests { /** * We leave the user IRQ off as much as possible, * but this means that requests will finish and never * be retired once the system goes idle. Set a timer to * fire periodically while the ring is running. When it * fires, go retire requests. */ struct delayed_work retire_work; } requests; struct { struct llist_head list; struct work_struct work; } watchdog; struct intel_wakeref wakeref; atomic_t user_wakeref; struct list_head closed_vma; spinlock_t closed_lock; /* guards the list of closed_vma */ ktime_t last_init_time; struct intel_reset reset; /** * Is the GPU currently considered idle, or busy executing * userspace requests? Whilst idle, we allow runtime power * management to power down the hardware and display clocks. * In order to reduce the effect on performance, there * is a slight delay before we do so. */ intel_wakeref_t awake; u32 clock_frequency; u32 clock_period_ns; struct intel_llc llc; struct intel_rc6 rc6; struct intel_rps rps; spinlock_t irq_lock; u32 gt_imr; u32 pm_ier; u32 pm_imr; u32 pm_guc_events; struct { bool active; /** * @lock: Lock protecting the below fields. */ seqcount_mutex_t lock; /** * @total: Total time this engine was busy. * * Accumulated time not counting the most recent block in cases * where engine is currently busy (active > 0). */ ktime_t total; /** * @start: Timestamp of the last idle to active transition. * * Idle is defined as active == 0, active is active > 0. */ ktime_t start; } stats; struct intel_engine_cs *engine[I915_NUM_ENGINES]; struct intel_engine_cs *engine_class[MAX_ENGINE_CLASS + 1] [MAX_ENGINE_INSTANCE + 1]; enum intel_submission_method submission_method; /* * Default address space (either GGTT or ppGTT depending on arch). * * Reserved for exclusive use by the kernel. */ struct i915_address_space *vm; /* * A pool of objects to use as shadow copies of client batch buffers * when the command parser is enabled. Prevents the client from * modifying the batch contents after software parsing. * * Buffers older than 1s are periodically reaped from the pool, * or may be reclaimed by the shrinker before then. */ struct intel_gt_buffer_pool buffer_pool; struct i915_vma *scratch; struct intel_migrate migrate; const struct intel_mmio_range *steering_table[NUM_STEERING_TYPES]; struct intel_gt_info { intel_engine_mask_t engine_mask; u32 l3bank_mask; u8 num_engines; /* General presence of SFC units */ u8 sfc_mask; /* Media engine access to SFC per instance */ u8 vdbox_sfc_access; /* Slice/subslice/EU info */ struct sseu_dev_info sseu; unsigned long mslice_mask; } info; struct { u8 uc_index; } mocs; struct intel_pxp pxp; }; enum intel_gt_scratch_field { /* 8 bytes */ INTEL_GT_SCRATCH_FIELD_DEFAULT = 0, /* 8 bytes */ INTEL_GT_SCRATCH_FIELD_RENDER_FLUSH = 128, /* 8 bytes */ INTEL_GT_SCRATCH_FIELD_COHERENTL3_WA = 256, /* 6 * 8 bytes */ INTEL_GT_SCRATCH_FIELD_PERF_CS_GPR = 2048, /* 4 bytes */ INTEL_GT_SCRATCH_FIELD_PERF_PREDICATE_RESULT_1 = 2096, }; #endif /* __INTEL_GT_TYPES_H__ */