|
@@ -32,7 +32,7 @@
|
|
|
"id": "5860ba9f-2db3-4480-b96b-4be1c68981eb",
|
|
"id": "5860ba9f-2db3-4480-b96b-4be1c68981eb",
|
|
|
"metadata": {},
|
|
"metadata": {},
|
|
|
"source": [
|
|
"source": [
|
|
|
- "We can pad the inputs to the maximum number of tokens to the maximum the model supports by setting the max length to\n",
|
|
|
|
|
|
|
+ "We can pad the inputs to the maximum number of tokens the model supports by setting the max length to 1024:\n",
|
|
|
"\n",
|
|
"\n",
|
|
|
"```python\n",
|
|
"```python\n",
|
|
|
"max_length = 1024\n",
|
|
"max_length = 1024\n",
|
|
@@ -112,7 +112,7 @@
|
|
|
"id": "7418a629-51b6-4aa2-83b7-bc0261bc370f",
|
|
"id": "7418a629-51b6-4aa2-83b7-bc0261bc370f",
|
|
|
"metadata": {},
|
|
"metadata": {},
|
|
|
"source": [
|
|
"source": [
|
|
|
- "ther than finetuning the last output token, we can finetune the first output token by changing \n",
|
|
|
|
|
|
|
+ "Rather than finetuning the last output token, we can finetune the first output token by changing \n",
|
|
|
"\n",
|
|
"\n",
|
|
|
"```python\n",
|
|
"```python\n",
|
|
|
"model(input_batch)[:, -1, :]\n",
|
|
"model(input_batch)[:, -1, :]\n",
|