@@ -5664,10 +5664,10 @@ void ggml_init_cublas() {
5664
5664
GGML_ASSERT (g_device_count <= GGML_CUDA_MAX_DEVICES);
5665
5665
int64_t total_vram = 0 ;
5666
5666
fprintf (stderr, " %s: found %d " GGML_CUDA_NAME " devices:\n " , __func__, g_device_count);
5667
- for (int64_t id = 0 ; id < g_device_count; ++id) {
5667
+ for (int id = 0 ; id < g_device_count; ++id) {
5668
5668
cudaDeviceProp prop;
5669
5669
CUDA_CHECK (cudaGetDeviceProperties (&prop, id));
5670
- fprintf (stderr, " Device %ld : %s, compute capability %d.%d\n " , id, prop.name , prop.major , prop.minor );
5670
+ fprintf (stderr, " Device %d : %s, compute capability %d.%d\n " , id, prop.name , prop.major , prop.minor );
5671
5671
5672
5672
g_tensor_split[id] = total_vram;
5673
5673
total_vram += prop.totalGlobalMem ;
@@ -5677,15 +5677,15 @@ void ggml_init_cublas() {
5677
5677
g_compute_capabilities[id] = 100 *prop.major + 10 *prop.minor ;
5678
5678
#endif // defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__)
5679
5679
}
5680
- for (int64_t id = 0 ; id < g_device_count; ++id) {
5680
+ for (int id = 0 ; id < g_device_count; ++id) {
5681
5681
g_tensor_split[id] /= total_vram;
5682
5682
}
5683
5683
5684
- for (int64_t id = 0 ; id < g_device_count; ++id) {
5684
+ for (int id = 0 ; id < g_device_count; ++id) {
5685
5685
CUDA_CHECK (ggml_cuda_set_device (id));
5686
5686
5687
5687
// create cuda streams
5688
- for (int64_t is = 0 ; is < MAX_STREAMS; ++is) {
5688
+ for (int is = 0 ; is < MAX_STREAMS; ++is) {
5689
5689
CUDA_CHECK (cudaStreamCreateWithFlags (&g_cudaStreams[id][is], cudaStreamNonBlocking));
5690
5690
}
5691
5691
0 commit comments