Skip to content

Commit

Permalink
Merge branch 'mini-hack-night' into v4
Browse files Browse the repository at this point in the history
  • Loading branch information
vintrocode committed Dec 14, 2023
2 parents e7b1cec + 8a5c8c9 commit 9196242
Show file tree
Hide file tree
Showing 28 changed files with 31 additions and 909 deletions.
1 change: 1 addition & 0 deletions content/blog/Open-Sourcing Tutor-GPT.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
---
title: "Open-Sourcing Tutor-GPT"
date: "Jun 2, 2023"
---

![[assets/human_machine_learning.jpeg]]
Expand Down
1 change: 1 addition & 0 deletions content/blog/Theory-of-Mind Is All You Need.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
---
title: "Theory-of-Mind Is All You Need"
date: "Jun 12, 2023"
---

## TL;DR
Expand Down
40 changes: 0 additions & 40 deletions content/notes/CJK + Latex Support (测试).md

This file was deleted.

Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
TL;DR: they aren't very flexible for intermediate metacognition steps

It's interesting that the machine learning community has decided to converge on this training paradigm because it assumes only two participants in a conversation. Just thinking intuitively about what happens when you train/fine-tune a language model, you being to reinforce token distributions that are appropriate to come in between the special tokens denoting human vs AI messages.

The issue we see here is that oftentimes there are a lot of intermediate reasoning steps you want to take in order to serve a more socially-aware answer. It's almost like the current state of inference is the equivalent of saying the first thing that comes to mind -- the quickness of one's wit can vary, but usually we think for a second before responding. We saw the advantages of doing this with Bloom (see [[Theory-of-Mind Is All You Need]]) and continue to be interested in exploring how much better this can get.

In order to assess its efficacy in this regard, I usually want to prompt it to generate as if it were the user -- which is usually very hard given the fact that those types of responses don't ever come after the special AI message token.

We're already anecdotally seeing very well-trained completion models follow instructions well likely because of its incorporation in their pre-training. Is chat the next thing to be subsumed by general completion models? Because if so, flexibility in the types of inferences you can make would be very beneficial. Metacognition becomes something you can do at any step in a conversation. Same with instruction following and chat. Maybe this is what starts to move language models in a much more general direction.
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
For wetware, metacognition is typically defined as "thinking about thinking" or often a catch-all for any "higher-order" cognition. In some more specific domains, it's an introspective process, focused on thinking about your own thinking.

In large language models, the synthetic corollary of cognition is inference. So we can reasonably call a metacognitive process in an LLM as any that runs inference on the result of prior inference. That is, inference itself is used as context. It might be instantly funneled into the next prompt, stored for later use, or leveraged by another model. Experiments here will be critical to overcome the machine learning community's fixation of task completion (see [[The machine learning industry is too focused on general task performance]]).
63 changes: 0 additions & 63 deletions content/notes/callouts.md

This file was deleted.

229 changes: 0 additions & 229 deletions content/notes/config.md

This file was deleted.

17 changes: 0 additions & 17 deletions content/notes/custom Domain.md

This file was deleted.

Loading

0 comments on commit 9196242

Please sign in to comment.