@@ -13,22 +13,22 @@ public unsafe struct LLamaModelParams
1313 /// number of layers to store in VRAM
1414 /// </summary>
1515 public int n_gpu_layers ;
16-
16+
1717 /// <summary>
1818 /// how to split the model across multiple GPUs
1919 /// </summary>
2020 public GPUSplitMode split_mode ;
21-
21+
2222 /// <summary>
2323 /// the GPU that is used for scratch and small tensors
2424 /// </summary>
2525 public int main_gpu ;
26-
26+
2727 /// <summary>
2828 /// how to split layers across multiple GPUs (size: <see cref="NativeApi.llama_max_devices"/>)
2929 /// </summary>
3030 public float * tensor_split ;
31-
31+
3232 /// <summary>
3333 /// called with a progress value between 0 and 1, pass NULL to disable. If the provided progress_callback
3434 /// returns true, model loading continues. If it returns false, model loading is immediately aborted.
@@ -40,17 +40,17 @@ public unsafe struct LLamaModelParams
4040#else
4141 public LlamaProgressCallback ? progress_callback ;
4242#endif
43-
43+
4444 /// <summary>
4545 /// context pointer passed to the progress callback
4646 /// </summary>
4747 public void * progress_callback_user_data ;
48-
48+
4949 /// <summary>
5050 /// override key-value pairs of the model meta data
5151 /// </summary>
5252 public LLamaModelMetadataOverride * kv_overrides ;
53-
53+
5454 /// <summary>
5555 /// only load the vocabulary, no weights
5656 /// </summary>
@@ -59,9 +59,9 @@ public bool vocab_only
5959 readonly get => Convert . ToBoolean ( _vocab_only ) ;
6060 set => _vocab_only = Convert . ToSByte ( value ) ;
6161 }
62-
62+
6363 private sbyte _vocab_only ;
64-
64+
6565 /// <summary>
6666 /// use mmap if possible
6767 /// </summary>
@@ -70,9 +70,9 @@ public bool use_mmap
7070 readonly get => Convert . ToBoolean ( _use_mmap ) ;
7171 set => _use_mmap = Convert . ToSByte ( value ) ;
7272 }
73-
73+
7474 private sbyte _use_mmap ;
75-
75+
7676 /// <summary>
7777 /// force system to keep model in RAM
7878 /// </summary>
@@ -81,17 +81,17 @@ public bool use_mlock
8181 readonly get => Convert . ToBoolean ( _use_mlock ) ;
8282 set => _use_mlock = Convert . ToSByte ( value ) ;
8383 }
84-
84+
8585 private sbyte _use_mlock ;
86-
86+
8787 /// <summary>
8888 /// Create a LLamaModelParams with default values
8989 /// </summary>
9090 /// <returns></returns>
9191 public static LLamaModelParams Default ( )
9292 {
9393 return llama_model_default_params ( ) ;
94-
94+
9595 [ DllImport ( NativeApi . libraryName , CallingConvention = CallingConvention . Cdecl ) ]
9696 static extern LLamaModelParams llama_model_default_params ( ) ;
9797 }
0 commit comments