diff --git a/LLama/Common/ModelParams.cs b/LLama/Common/ModelParams.cs index 6d19a8de9..f1cef0723 100644 --- a/LLama/Common/ModelParams.cs +++ b/LLama/Common/ModelParams.cs @@ -111,54 +111,6 @@ private ModelParams() // This constructor (default parameterless constructor) is used by Newtonsoft to deserialize! ModelPath = ""; } - - /// - /// - /// - /// The model path. - /// Model context size (n_ctx) - /// Number of layers to run in VRAM / GPU memory (n_gpu_layers) - /// Seed for the random number generator (seed) - /// Whether to use f16 instead of f32 for memory kv (memory_f16) - /// Whether to use mmap for faster loads (use_mmap) - /// Whether to use mlock to keep model in memory (use_mlock) - /// Thether to compute perplexity over the prompt (perplexity) - /// Lora adapter path (lora_adapter) - /// Base model path for the lora adapter (lora_base) - /// Number of threads (-1 = autodetect) (n_threads) - /// Batch size for prompt processing (must be >=32 to use BLAS) (n_batch) - /// Whether to use embedding mode. (embedding) Note that if this is set to true, The LLamaModel won't produce text response anymore. - /// RoPE base frequency. - /// RoPE frequency scaling factor - /// Use experimental mul_mat_q kernels - /// The encoding to use to convert text for the model - [Obsolete("Use object initializer to set all optional parameters")] - public ModelParams(string modelPath, uint contextSize = 512, int gpuLayerCount = 20, - uint seed = 1337, bool useFp16Memory = true, - bool useMemorymap = true, bool useMemoryLock = false, bool perplexity = false, - string loraAdapter = "", string loraBase = "", int threads = -1, uint batchSize = 512, - bool embeddingMode = false, - float? ropeFrequencyBase = null, float? ropeFrequencyScale = null, bool mulMatQ = false, - string encoding = "UTF-8") - { - ContextSize = contextSize; - GpuLayerCount = gpuLayerCount; - Seed = seed; - UseFp16Memory = useFp16Memory; - UseMemorymap = useMemorymap; - UseMemoryLock = useMemoryLock; - Perplexity = perplexity; - ModelPath = modelPath; - LoraBase = loraBase; - Threads = threads < 1 ? null : (uint)threads; - BatchSize = batchSize; - EmbeddingMode = embeddingMode; - RopeFrequencyBase = ropeFrequencyBase; - RopeFrequencyScale = ropeFrequencyScale; - MulMatQ = mulMatQ; - Encoding = Encoding.GetEncoding(encoding); - LoraAdapters.Add(new LoraAdapter(loraAdapter, 1)); - } } internal class EncodingConverter