Skip to content

Commit

Permalink
nbstripout
Browse files Browse the repository at this point in the history
  • Loading branch information
Huizerd committed Sep 5, 2024
1 parent 42c7128 commit c4f4c8c
Showing 1 changed file with 4 additions and 18 deletions.
22 changes: 4 additions & 18 deletions ex_1/ex_1.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@
},
{
"cell_type": "code",
"execution_count": 2,
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
Expand Down Expand Up @@ -166,7 +166,7 @@
},
{
"cell_type": "code",
"execution_count": 3,
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
Expand Down Expand Up @@ -366,23 +366,9 @@
},
{
"cell_type": "code",
"execution_count": 4,
"execution_count": null,
"metadata": {},
"outputs": [
{
"ename": "FileNotFoundError",
"evalue": "[Errno 2] No such file or directory: '/workspaces/msc-ai-course/data/2D_QUAD_HOVER.npz'",
"output_type": "error",
"traceback": [
"\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
"\u001b[0;31mFileNotFoundError\u001b[0m Traceback (most recent call last)",
"Cell \u001b[0;32mIn[4], line 1\u001b[0m\n\u001b[0;32m----> 1\u001b[0m train_set, val_set \u001b[38;5;241m=\u001b[39m prepare_dataset(config[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mdataset_file\u001b[39m\u001b[38;5;124m\"\u001b[39m], config[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mtrajectory_dim\u001b[39m\u001b[38;5;124m\"\u001b[39m])\n",
"Cell \u001b[0;32mIn[3], line 135\u001b[0m, in \u001b[0;36mprepare_dataset\u001b[0;34m(file_path, d)\u001b[0m\n\u001b[1;32m 132\u001b[0m \u001b[38;5;28;01massert\u001b[39;00m d \u001b[38;5;129;01min\u001b[39;00m [\u001b[38;5;241m2\u001b[39m, \u001b[38;5;241m3\u001b[39m], \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mdimension must be 2 or 3\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 134\u001b[0m \u001b[38;5;66;03m# Load the dataset\u001b[39;00m\n\u001b[0;32m--> 135\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m np\u001b[38;5;241m.\u001b[39mload(file_path) \u001b[38;5;28;01mas\u001b[39;00m full_dataset:\n\u001b[1;32m 136\u001b[0m \u001b[38;5;66;03m# Total number of trajectories\u001b[39;00m\n\u001b[1;32m 137\u001b[0m num \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mlen\u001b[39m(full_dataset[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mt\u001b[39m\u001b[38;5;124m\"\u001b[39m])\n\u001b[1;32m 139\u001b[0m \u001b[38;5;66;03m# Split dataset into training and validation sets\u001b[39;00m\n",
"File \u001b[0;32m/opt/conda/envs/AE4353/lib/python3.11/site-packages/numpy/lib/npyio.py:427\u001b[0m, in \u001b[0;36mload\u001b[0;34m(file, mmap_mode, allow_pickle, fix_imports, encoding, max_header_size)\u001b[0m\n\u001b[1;32m 425\u001b[0m own_fid \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mFalse\u001b[39;00m\n\u001b[1;32m 426\u001b[0m \u001b[38;5;28;01melse\u001b[39;00m:\n\u001b[0;32m--> 427\u001b[0m fid \u001b[38;5;241m=\u001b[39m stack\u001b[38;5;241m.\u001b[39menter_context(\u001b[38;5;28mopen\u001b[39m(os_fspath(file), \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mrb\u001b[39m\u001b[38;5;124m\"\u001b[39m))\n\u001b[1;32m 428\u001b[0m own_fid \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mTrue\u001b[39;00m\n\u001b[1;32m 430\u001b[0m \u001b[38;5;66;03m# Code to distinguish from NumPy binary files and pickles.\u001b[39;00m\n",
"\u001b[0;31mFileNotFoundError\u001b[0m: [Errno 2] No such file or directory: '/workspaces/msc-ai-course/data/2D_QUAD_HOVER.npz'"
]
}
],
"outputs": [],
"source": [
"train_set, val_set = prepare_dataset(config[\"dataset_file\"], config[\"trajectory_dim\"])"
]
Expand Down

0 comments on commit c4f4c8c

Please sign in to comment.