From 28595ead233183641fb47274311b86ad21bfbb8d Mon Sep 17 00:00:00 2001 From: = <=> Date: Tue, 21 Jan 2025 22:31:27 +1100 Subject: [PATCH] production typos --- src/content/first_ai_blog.md | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git a/src/content/first_ai_blog.md b/src/content/first_ai_blog.md index 98362b3..30413e7 100644 --- a/src/content/first_ai_blog.md +++ b/src/content/first_ai_blog.md @@ -12,11 +12,11 @@ As mentioned in the last post I have been experimenting with AI content generati It's still not in the state I want it to be and isn't connecting to the actual blog repo yet (the idea being that I edit and change as part of a review process). But it is generating stuff that I thought would be worth a share just for fun. The eventual idea is the container will be something I trigger as part of a CI/CD process when the equipment is up and running but the final "production" implementation is still being fleshed out in my head (if you can call a homelab project "production") -The focus to this point has been on prompt engineering and model selection. A big part of this is that it needs to be able to run completely indepdantly of any cloud services so no Chat GPT. +The focus to this point has been on prompt engineering and model selection. A big part of this is that it needs to be able to run completely independently of any cloud services so no Chat GPT. -The obvious solution is [ollama](https://ollama.com) I'm luck enough to have a modest secondary gaming rig in my living room with an nvidia 2060 in it that can act as a modest AI server so I have set it up there. +The obvious solution is [ollama](https://ollama.com) I'm lucky enough to have a modest secondary gaming rig in my living room with an nvidia 2060 in it that can act as a modest AI server so I have set it up there. -This server has allowed me to experiment almost at will with models. a few I tried included +This server has allowed me to experiment, almost at will, with models. a few I tried included - llama3.2 - zephyr @@ -29,6 +29,7 @@ However I have found that mistral-nemo seems to be the strongest contender, at l You can see the code and what I have been working on for more details [HERE](https://git.aridgwayweb.com/armistace/blog_creator) #### The summary prompt used by mistral to generate this post + _The following is what I have in trilium to generate the AI written content_ Get a python file that can communicate with a local ollama instance @@ -50,7 +51,8 @@ Send a notification to matrix that there is a new PR in and give a summary of th _as you can see it pretty light on content so what it generates given this lack of context I feel is pretty good_ # Let the post begin (begin...begin...begin...) -_What follows is what it generated.. yes it is hilarious but well... it was to good not to share_ + +_What follows is what it generated.. yes it is hilarious but well... it was too good not to share_ # G'day, mates! Let's make an Ollama Blog Writer!