chatGPT Skills

prompting is essential. good examples are collected here:

from a more systematic prompt engineer vedio:

code generation sample 2

GPT-4 compared to chatGPT is particularly better at reasoning, the prompt sample is

Next, moving to more advanced prompt engineering:

in context learning

CoT, chain of thought prompting

Zero-shot prompting involve adding “let’s think step by step”.

Self-consistency has an interesting example on age reasoning:

After applying smart/self-consistency prompting, the answer can be improved:

Generate knowledge prompting:

program-aided language model such as chain of thought prompting is to steer models to perform better at complex reasoning tasks, it can be augmented by program-aided Language models to generate intermediate reasoning steps.

Further combining: ReAct

prompt include demoing how to solve problem such as an extreme case like below:

For real applications which would explode in near future, data-augmented generation would be it!

Leave a comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.