@@ -974,8 +974,8 @@ bool CommandListCoreFamilyImmediate<gfxCoreFamily>::preferCopyThroughLockedPtr(C
974
974
bool cpuMemCopyEnabled = false ;
975
975
976
976
switch (transferType) {
977
- case HOST_USM_TO_DEVICE_USM :
978
- case DEVICE_USM_TO_HOST_USM : {
977
+ case TransferType::hostUsmToDeviceUsm :
978
+ case TransferType::deviceUsmToHostUsm : {
979
979
if (this ->dependenciesPresent ) {
980
980
cpuMemCopyEnabled = false ;
981
981
break ;
@@ -990,8 +990,8 @@ bool CommandListCoreFamilyImmediate<gfxCoreFamily>::preferCopyThroughLockedPtr(C
990
990
cpuMemCopyEnabled = allEventsCompleted;
991
991
break ;
992
992
}
993
- case HOST_NON_USM_TO_DEVICE_USM :
994
- case DEVICE_USM_TO_HOST_NON_USM :
993
+ case TransferType::hostNonUsmToDeviceUsm :
994
+ case TransferType::deviceUsmToHostNonUsm :
995
995
cpuMemCopyEnabled = true ;
996
996
break ;
997
997
default :
@@ -1004,19 +1004,19 @@ bool CommandListCoreFamilyImmediate<gfxCoreFamily>::preferCopyThroughLockedPtr(C
1004
1004
1005
1005
template <GFXCORE_FAMILY gfxCoreFamily>
1006
1006
bool CommandListCoreFamilyImmediate<gfxCoreFamily>::isSuitableUSMHostAlloc(NEO::SvmAllocationData *alloc) {
1007
- return alloc && (alloc->memoryType == InternalMemoryType::HOST_UNIFIED_MEMORY );
1007
+ return alloc && (alloc->memoryType == InternalMemoryType::hostUnifiedMemory );
1008
1008
}
1009
1009
1010
1010
template <GFXCORE_FAMILY gfxCoreFamily>
1011
1011
bool CommandListCoreFamilyImmediate<gfxCoreFamily>::isSuitableUSMDeviceAlloc(NEO::SvmAllocationData *alloc) {
1012
- return alloc && (alloc->memoryType == InternalMemoryType::DEVICE_UNIFIED_MEMORY ) &&
1012
+ return alloc && (alloc->memoryType == InternalMemoryType::deviceUnifiedMemory ) &&
1013
1013
alloc->gpuAllocations .getGraphicsAllocation (this ->device ->getRootDeviceIndex ()) &&
1014
1014
alloc->gpuAllocations .getGraphicsAllocation (this ->device ->getRootDeviceIndex ())->storageInfo .getNumBanks () == 1 ;
1015
1015
}
1016
1016
1017
1017
template <GFXCORE_FAMILY gfxCoreFamily>
1018
1018
bool CommandListCoreFamilyImmediate<gfxCoreFamily>::isSuitableUSMSharedAlloc(NEO::SvmAllocationData *alloc) {
1019
- return alloc && (alloc->memoryType == InternalMemoryType::SHARED_UNIFIED_MEMORY );
1019
+ return alloc && (alloc->memoryType == InternalMemoryType::sharedUnifiedMemory );
1020
1020
}
1021
1021
1022
1022
template <GFXCORE_FAMILY gfxCoreFamily>
@@ -1138,111 +1138,111 @@ TransferType CommandListCoreFamilyImmediate<gfxCoreFamily>::getTransferType(cons
1138
1138
const bool dstHostNonUSM = (cpuMemCopyInfo.dstAllocData == nullptr ) && !cpuMemCopyInfo.dstIsImportedHostPtr ;
1139
1139
1140
1140
if (srcHostNonUSM && dstHostUSM) {
1141
- return HOST_NON_USM_TO_HOST_USM ;
1141
+ return TransferType::hostNonUsmToHostUsm ;
1142
1142
}
1143
1143
if (srcHostNonUSM && dstDeviceUSM) {
1144
- return HOST_NON_USM_TO_DEVICE_USM ;
1144
+ return TransferType::hostNonUsmToDeviceUsm ;
1145
1145
}
1146
1146
if (srcHostNonUSM && dstSharedUSM) {
1147
- return HOST_NON_USM_TO_SHARED_USM ;
1147
+ return TransferType::hostNonUsmToSharedUsm ;
1148
1148
}
1149
1149
if (srcHostNonUSM && dstHostNonUSM) {
1150
- return HOST_NON_USM_TO_HOST_NON_USM ;
1150
+ return TransferType::hostNonUsmToHostNonUsm ;
1151
1151
}
1152
1152
1153
1153
if (srcHostUSM && dstHostUSM) {
1154
- return HOST_USM_TO_HOST_USM ;
1154
+ return TransferType::hostUsmToHostUsm ;
1155
1155
}
1156
1156
if (srcHostUSM && dstDeviceUSM) {
1157
- return HOST_USM_TO_DEVICE_USM ;
1157
+ return TransferType::hostUsmToDeviceUsm ;
1158
1158
}
1159
1159
if (srcHostUSM && dstSharedUSM) {
1160
- return HOST_USM_TO_SHARED_USM ;
1160
+ return TransferType::hostUsmToSharedUsm ;
1161
1161
}
1162
1162
if (srcHostUSM && dstHostNonUSM) {
1163
- return HOST_USM_TO_HOST_NON_USM ;
1163
+ return TransferType::hostUsmToHostNonUsm ;
1164
1164
}
1165
1165
1166
1166
if (srcDeviceUSM && dstHostUSM) {
1167
- return DEVICE_USM_TO_HOST_USM ;
1167
+ return TransferType::deviceUsmToHostUsm ;
1168
1168
}
1169
1169
if (srcDeviceUSM && dstDeviceUSM) {
1170
- return DEVICE_USM_TO_DEVICE_USM ;
1170
+ return TransferType::deviceUsmToDeviceUsm ;
1171
1171
}
1172
1172
if (srcDeviceUSM && dstSharedUSM) {
1173
- return DEVICE_USM_TO_SHARED_USM ;
1173
+ return TransferType::deviceUsmToSharedUsm ;
1174
1174
}
1175
1175
if (srcDeviceUSM && dstHostNonUSM) {
1176
- return DEVICE_USM_TO_HOST_NON_USM ;
1176
+ return TransferType::deviceUsmToHostNonUsm ;
1177
1177
}
1178
1178
1179
1179
if (srcSharedUSM && dstHostUSM) {
1180
- return SHARED_USM_TO_HOST_USM ;
1180
+ return TransferType::sharedUsmToHostUsm ;
1181
1181
}
1182
1182
if (srcSharedUSM && dstDeviceUSM) {
1183
- return SHARED_USM_TO_DEVICE_USM ;
1183
+ return TransferType::sharedUsmToDeviceUsm ;
1184
1184
}
1185
1185
if (srcSharedUSM && dstSharedUSM) {
1186
- return SHARED_USM_TO_SHARED_USM ;
1186
+ return TransferType::sharedUsmToSharedUsm ;
1187
1187
}
1188
1188
if (srcSharedUSM && dstHostNonUSM) {
1189
- return SHARED_USM_TO_HOST_NON_USM ;
1189
+ return TransferType::sharedUsmToHostNonUsm ;
1190
1190
}
1191
1191
1192
- return TRANSFER_TYPE_UNKNOWN ;
1192
+ return TransferType::unknown ;
1193
1193
}
1194
1194
1195
1195
template <GFXCORE_FAMILY gfxCoreFamily>
1196
1196
size_t CommandListCoreFamilyImmediate<gfxCoreFamily>::getTransferThreshold(TransferType transferType) {
1197
1197
size_t retVal = 0u ;
1198
1198
1199
1199
switch (transferType) {
1200
- case HOST_NON_USM_TO_HOST_USM :
1200
+ case TransferType::hostNonUsmToHostUsm :
1201
1201
retVal = 1 * MemoryConstants::megaByte;
1202
1202
break ;
1203
- case HOST_NON_USM_TO_DEVICE_USM :
1203
+ case TransferType::hostNonUsmToDeviceUsm :
1204
1204
retVal = 4 * MemoryConstants::megaByte;
1205
1205
if (NEO::debugManager.flags .ExperimentalH2DCpuCopyThreshold .get () != -1 ) {
1206
1206
retVal = NEO::debugManager.flags .ExperimentalH2DCpuCopyThreshold .get ();
1207
1207
}
1208
1208
break ;
1209
- case HOST_NON_USM_TO_SHARED_USM :
1209
+ case TransferType::hostNonUsmToSharedUsm :
1210
1210
retVal = 0u ;
1211
1211
break ;
1212
- case HOST_NON_USM_TO_HOST_NON_USM :
1212
+ case TransferType::hostNonUsmToHostNonUsm :
1213
1213
retVal = 1 * MemoryConstants::megaByte;
1214
1214
break ;
1215
- case HOST_USM_TO_HOST_USM :
1215
+ case TransferType::hostUsmToHostUsm :
1216
1216
retVal = 200 * MemoryConstants::kiloByte;
1217
1217
break ;
1218
- case HOST_USM_TO_DEVICE_USM :
1218
+ case TransferType::hostUsmToDeviceUsm :
1219
1219
retVal = 50 * MemoryConstants::kiloByte;
1220
1220
break ;
1221
- case HOST_USM_TO_SHARED_USM :
1221
+ case TransferType::hostUsmToSharedUsm :
1222
1222
retVal = 0u ;
1223
1223
break ;
1224
- case HOST_USM_TO_HOST_NON_USM :
1224
+ case TransferType::hostUsmToHostNonUsm :
1225
1225
retVal = 500 * MemoryConstants::kiloByte;
1226
1226
break ;
1227
- case DEVICE_USM_TO_DEVICE_USM :
1227
+ case TransferType::deviceUsmToDeviceUsm :
1228
1228
retVal = 0u ;
1229
1229
break ;
1230
- case DEVICE_USM_TO_SHARED_USM :
1230
+ case TransferType::deviceUsmToSharedUsm :
1231
1231
retVal = 0u ;
1232
1232
break ;
1233
- case DEVICE_USM_TO_HOST_USM :
1233
+ case TransferType::deviceUsmToHostUsm :
1234
1234
retVal = 128u ;
1235
1235
break ;
1236
- case DEVICE_USM_TO_HOST_NON_USM :
1236
+ case TransferType::deviceUsmToHostNonUsm :
1237
1237
retVal = 1 * MemoryConstants::kiloByte;
1238
1238
if (NEO::debugManager.flags .ExperimentalD2HCpuCopyThreshold .get () != -1 ) {
1239
1239
retVal = NEO::debugManager.flags .ExperimentalD2HCpuCopyThreshold .get ();
1240
1240
}
1241
1241
break ;
1242
- case SHARED_USM_TO_HOST_USM :
1243
- case SHARED_USM_TO_DEVICE_USM :
1244
- case SHARED_USM_TO_SHARED_USM :
1245
- case SHARED_USM_TO_HOST_NON_USM :
1242
+ case TransferType::sharedUsmToHostUsm :
1243
+ case TransferType::sharedUsmToDeviceUsm :
1244
+ case TransferType::sharedUsmToSharedUsm :
1245
+ case TransferType::sharedUsmToHostNonUsm :
1246
1246
retVal = 0u ;
1247
1247
break ;
1248
1248
default :
0 commit comments