<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:media="http://search.yahoo.com/mrss/"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>gguf - 四号程序员</title>
	<atom:link href="https://www.coder4.com/archives/tag/gguf/feed" rel="self" type="application/rss+xml" />
	<link>https://www.coder4.com</link>
	<description>Keep It Simple and Stupid</description>
	<lastBuildDate>Wed, 12 Feb 2025 04:47:21 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.8.3</generator>
	<item>
		<title>llama.cpp跑gguf量化模型</title>
		<link>https://www.coder4.com/archives/8276</link>
					<comments>https://www.coder4.com/archives/8276#respond</comments>
		
		<dc:creator><![CDATA[coder4]]></dc:creator>
		<pubDate>Tue, 11 Feb 2025 11:19:37 +0000</pubDate>
				<category><![CDATA[Linux]]></category>
		<category><![CDATA[gguf]]></category>
		<category><![CDATA[llama.cpp]]></category>
		<guid isPermaLink="false">https://www.coder4.com/?p=8276</guid>

					<description><![CDATA[gguf量化的模型，目前看只有llama.cpp支持的比较好，使用了ollama和其他衍生客户端，都会有奇怪的乱输出问题，还没找到解法（据说是要配置template）。 首先这里下载，https://github.com/ggerganov/llama.cpp/releases 解压缩 unzip llama-b4686-bin-ubuntu-x64.zip 运行 ./llama-b4686-bin-ubuntu-x64/build/bin/llama-cli -m ./xxx[......] 继续阅读]]></description>
		
					<wfw:commentRss>https://www.coder4.com/archives/8276/feed</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
