@@ -321,7 +321,7 @@ static int eb_pin_engine(struct i915_execbuffer *eb, bool throttle);
321
321
static void eb_unpin_engine (struct i915_execbuffer * eb );
322
322
static void eb_capture_release (struct i915_execbuffer * eb );
323
323
324
- static inline bool eb_use_cmdparser (const struct i915_execbuffer * eb )
324
+ static bool eb_use_cmdparser (const struct i915_execbuffer * eb )
325
325
{
326
326
return intel_engine_requires_cmd_parser (eb -> context -> engine ) ||
327
327
(intel_engine_using_cmd_parser (eb -> context -> engine ) &&
@@ -433,7 +433,7 @@ static u64 eb_pin_flags(const struct drm_i915_gem_exec_object2 *entry,
433
433
return pin_flags ;
434
434
}
435
435
436
- static inline int
436
+ static int
437
437
eb_pin_vma (struct i915_execbuffer * eb ,
438
438
const struct drm_i915_gem_exec_object2 * entry ,
439
439
struct eb_vma * ev )
@@ -486,7 +486,7 @@ eb_pin_vma(struct i915_execbuffer *eb,
486
486
return 0 ;
487
487
}
488
488
489
- static inline void
489
+ static void
490
490
eb_unreserve_vma (struct eb_vma * ev )
491
491
{
492
492
if (unlikely (ev -> flags & __EXEC_OBJECT_HAS_FENCE ))
@@ -548,7 +548,7 @@ eb_validate_vma(struct i915_execbuffer *eb,
548
548
return 0 ;
549
549
}
550
550
551
- static inline bool
551
+ static bool
552
552
is_batch_buffer (struct i915_execbuffer * eb , unsigned int buffer_idx )
553
553
{
554
554
return eb -> args -> flags & I915_EXEC_BATCH_FIRST ?
@@ -628,8 +628,8 @@ eb_add_vma(struct i915_execbuffer *eb,
628
628
return 0 ;
629
629
}
630
630
631
- static inline int use_cpu_reloc (const struct reloc_cache * cache ,
632
- const struct drm_i915_gem_object * obj )
631
+ static int use_cpu_reloc (const struct reloc_cache * cache ,
632
+ const struct drm_i915_gem_object * obj )
633
633
{
634
634
if (!i915_gem_object_has_struct_page (obj ))
635
635
return false;
@@ -1107,7 +1107,7 @@ static void eb_destroy(const struct i915_execbuffer *eb)
1107
1107
kfree (eb -> buckets );
1108
1108
}
1109
1109
1110
- static inline u64
1110
+ static u64
1111
1111
relocation_target (const struct drm_i915_gem_relocation_entry * reloc ,
1112
1112
const struct i915_vma * target )
1113
1113
{
@@ -1128,19 +1128,19 @@ static void reloc_cache_init(struct reloc_cache *cache,
1128
1128
cache -> node .flags = 0 ;
1129
1129
}
1130
1130
1131
- static inline void * unmask_page (unsigned long p )
1131
+ static void * unmask_page (unsigned long p )
1132
1132
{
1133
1133
return (void * )(uintptr_t )(p & PAGE_MASK );
1134
1134
}
1135
1135
1136
- static inline unsigned int unmask_flags (unsigned long p )
1136
+ static unsigned int unmask_flags (unsigned long p )
1137
1137
{
1138
1138
return p & ~PAGE_MASK ;
1139
1139
}
1140
1140
1141
1141
#define KMAP 0x4 /* after CLFLUSH_FLAGS */
1142
1142
1143
- static inline struct i915_ggtt * cache_to_ggtt (struct reloc_cache * cache )
1143
+ static struct i915_ggtt * cache_to_ggtt (struct reloc_cache * cache )
1144
1144
{
1145
1145
struct drm_i915_private * i915 =
1146
1146
container_of (cache , struct i915_execbuffer , reloc_cache )-> i915 ;
0 commit comments