Lines Matching refs:read
209 pm_needed = ss->read.size > XE_DEVCOREDUMP_CHUNK_MAX; in xe_devcoredump_read()
215 if (!ss->read.buffer) { in xe_devcoredump_read()
220 if (offset >= ss->read.size) in xe_devcoredump_read()
227 if (offset >= ss->read.chunk_position + XE_DEVCOREDUMP_CHUNK_MAX || in xe_devcoredump_read()
228 offset < ss->read.chunk_position) { in xe_devcoredump_read()
229 ss->read.chunk_position = new_chunk_position * in xe_devcoredump_read()
232 __xe_devcoredump_read(ss->read.buffer, in xe_devcoredump_read()
234 ss->read.chunk_position, coredump); in xe_devcoredump_read()
237 byte_copied = count < ss->read.size - offset ? count : in xe_devcoredump_read()
238 ss->read.size - offset; in xe_devcoredump_read()
239 memcpy(buffer, ss->read.buffer + chunk_offset, byte_copied); in xe_devcoredump_read()
263 kvfree(coredump->snapshot.read.buffer); in xe_devcoredump_free()
300 ss->read.chunk_position = 0; in xe_devcoredump_deferred_snap_work()
303 ss->read.size = __xe_devcoredump_read(NULL, LONG_MAX, 0, coredump); in xe_devcoredump_deferred_snap_work()
305 if (ss->read.size > XE_DEVCOREDUMP_CHUNK_MAX) { in xe_devcoredump_deferred_snap_work()
306 ss->read.buffer = kvmalloc(XE_DEVCOREDUMP_CHUNK_MAX, in xe_devcoredump_deferred_snap_work()
308 if (!ss->read.buffer) in xe_devcoredump_deferred_snap_work()
311 __xe_devcoredump_read(ss->read.buffer, in xe_devcoredump_deferred_snap_work()
315 ss->read.buffer = kvmalloc(ss->read.size, GFP_USER); in xe_devcoredump_deferred_snap_work()
316 if (!ss->read.buffer) in xe_devcoredump_deferred_snap_work()
319 __xe_devcoredump_read(ss->read.buffer, ss->read.size, 0, in xe_devcoredump_deferred_snap_work()