Conversation about chatGPT

We had a nice conversation about chatGPT, the pluses, uses, minuses, concerns, alternatives, and all kinds of stuff.

Miro board with notes: Miro | Online Whiteboard for Visual Collaboration

We talked about the importance of the training data, and value in it. Here’s how chatGPT handles training data, with some concrete examples: OpenAI API

I recorded most of it with Otter, and then fed it to chatGPT to summarize (just to try it out). It did a bad job… couldn’t really take all of the discussion because it was too long (too many tokens!!!). I also can’t upload the raw transcript either :\ But here’s the links to the transcripts (had to make it into two sets):


Also, just want to note that in their docs you cannot (even with paid plans) increase the total number of tokens per request (which is 4096 for the fanciest model) - OpenAI API.

You can improve the quality of your custom model by fine tuning it (OpenAI API). Fine tuning gives it query - answer pairs to provide more accurate answers with. However, you can’t really just feed it a bunch of new bulk text (like all my meeting notes or something) and have it ‘learn’ from that the same way it ‘learned’ from the internet. Rather, you’d have to feed it only < 4096 tokens worth (which is not that much, a page or so).

So basically, we can’t really make a model for our local selves / groups per se, because that seems (at least how chatGPT is set up now) it would require a more significant retraining.

We can however train it to do specific things… like I’m imagining would be really useful for our community, training data that said: `input: , output: … this could be a farmer talking about his or her actions during that day converted to a farm management data format (JSON schema maybe) as an example.

This just reminded me of our conversation:

Was it Vic who mentioned a similar instance of getting a recommendation for a scholarly article that didn’t exist?

NYT just did a few pieces on the Bing chatbot. This is the long transcript of the conversation :
Follow-up articles were also interesting but this is the moment in time our future historians will be talking about… human or otherwise.

We discussed the possibility of replacing apis or other interactions that have the be very manually plumbed…

1 Like

Interesting newsletter on “open-ish-ness” in ML, and more broadly:

Reminds me, too, of @brianwdavis’s post:


1 Like