27
RPG daily blog summaries at dndblogs.com
(dndblogs.com)
This community is for meaningful discussions of tabletop/pen & paper RPGs
Rules (wip):
Yeah, GPT-4 is way more costly. It's rolled out to the paid tier of users AFAIK, though, so in your script it's as easy as subbing out "3.5-turbo" with "4" -- it may be worth just doing some trial runs and seeing if it's markedly improved. I expect though there's some room for tweaks with the 3.5-turbo model, though.
Temperature is something you can play with to get more predictable or more creative results; and is set between 0 and 1 with .1/.2 being more useful for every wrote data entry stuff and .7-.9 being more creative. That's something to play with (I'd probably start around .6 or .7). But tweaking the prompt itself and the "role:" instructions will almost certainly return the most immediate benefits.
I'm assuming your code looks something like:
response = openai.ChatCompletion.create( model="gpt-3.5-turbo", temperature=0.2, messages=[ { "role": "user", "content": f"Please summarize this blog post in 3-5 bullet points: {blog}" }
You can add a "role": "system" instruction to make clear that, e.g., the model is supposed to assume that they are summarizing this work for a dm/ttrpg audience, and reiterate that in your user content message as well:
response = openai.ChatCompletion.create( model="gpt-3.5-turbo-16k", temperature=0.6, messages=[ { "role": "system", "content": "You are an assistant for a TTRPG blog designed to help gamemasters and dungeonmasters. You have been tasked with reading blog posts from other game masters and identifying the creative and interesting takeaways and summarizing them for a DM/GM audience. The factual summarizing of the structure of the blog is less important than the relevant actionable advice for running interesting and engaging TTRPG sessions" }, { "role": "user", "content": f"Please summarize this ttrpg blog post in 3-5 bullet points for an audience of gamemasters and ttrpg enthusiasts, identifying the most interesting, actionable and creative take-aways: {blog}" }
Go crazy with those instructions, though. That's the real secret to getting good results from GPT chat/completion models. I also swapped out the vanilla 3.5-turbo model with the only barely more expensive 'turbo-16k' model, which has a higher token limit, and bumped the temperature up a bit.
Great stuff! I'll experiment with it and see how I can get better results. I've already been using the turbo-16k so I can feed large blog articles back to it. I also need to find a way to limit the total tokens coming out to ensure the summaries don't get to long sometimes.
Awesome! Looking forward to checking it out!! I believe there is a max_tokens flag you can add, but I haven't futzed with that https://platform.openai.com/docs/api-reference/chat/create#chat/create-max_tokens