Skip to content

Commit 9c314b4

Browse files
committed
Format file.
1 parent d7231f3 commit 9c314b4

File tree

1 file changed

+14
-14
lines changed

1 file changed

+14
-14
lines changed

LLama/Native/LLamaModelParams.cs

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -13,22 +13,22 @@ public unsafe struct LLamaModelParams
1313
/// number of layers to store in VRAM
1414
/// </summary>
1515
public int n_gpu_layers;
16-
16+
1717
/// <summary>
1818
/// how to split the model across multiple GPUs
1919
/// </summary>
2020
public GPUSplitMode split_mode;
21-
21+
2222
/// <summary>
2323
/// the GPU that is used for scratch and small tensors
2424
/// </summary>
2525
public int main_gpu;
26-
26+
2727
/// <summary>
2828
/// how to split layers across multiple GPUs (size: <see cref="NativeApi.llama_max_devices"/>)
2929
/// </summary>
3030
public float* tensor_split;
31-
31+
3232
/// <summary>
3333
/// called with a progress value between 0 and 1, pass NULL to disable. If the provided progress_callback
3434
/// returns true, model loading continues. If it returns false, model loading is immediately aborted.
@@ -40,17 +40,17 @@ public unsafe struct LLamaModelParams
4040
#else
4141
public LlamaProgressCallback? progress_callback;
4242
#endif
43-
43+
4444
/// <summary>
4545
/// context pointer passed to the progress callback
4646
/// </summary>
4747
public void* progress_callback_user_data;
48-
48+
4949
/// <summary>
5050
/// override key-value pairs of the model meta data
5151
/// </summary>
5252
public LLamaModelMetadataOverride* kv_overrides;
53-
53+
5454
/// <summary>
5555
/// only load the vocabulary, no weights
5656
/// </summary>
@@ -59,9 +59,9 @@ public bool vocab_only
5959
readonly get => Convert.ToBoolean(_vocab_only);
6060
set => _vocab_only = Convert.ToSByte(value);
6161
}
62-
62+
6363
private sbyte _vocab_only;
64-
64+
6565
/// <summary>
6666
/// use mmap if possible
6767
/// </summary>
@@ -70,9 +70,9 @@ public bool use_mmap
7070
readonly get => Convert.ToBoolean(_use_mmap);
7171
set => _use_mmap = Convert.ToSByte(value);
7272
}
73-
73+
7474
private sbyte _use_mmap;
75-
75+
7676
/// <summary>
7777
/// force system to keep model in RAM
7878
/// </summary>
@@ -81,17 +81,17 @@ public bool use_mlock
8181
readonly get => Convert.ToBoolean(_use_mlock);
8282
set => _use_mlock = Convert.ToSByte(value);
8383
}
84-
84+
8585
private sbyte _use_mlock;
86-
86+
8787
/// <summary>
8888
/// Create a LLamaModelParams with default values
8989
/// </summary>
9090
/// <returns></returns>
9191
public static LLamaModelParams Default()
9292
{
9393
return llama_model_default_params();
94-
94+
9595
[DllImport(NativeApi.libraryName, CallingConvention = CallingConvention.Cdecl)]
9696
static extern LLamaModelParams llama_model_default_params();
9797
}

0 commit comments

Comments
 (0)