Upload 10 files
Browse files- Copy of andrea-dd-dataLte100.npy +3 -0
- README.md +2 -0
- classicml.ipynb +0 -0
- graphformers (1).ipynb +0 -0
- proj2.ipynb +843 -0
- requirements.txt +91 -0
- san_best.pth +3 -0
- test_classicml.ipynb +0 -0
- transformer_classifier_relu.pth +3 -0
- tsne_data.csv +0 -0
Copy of andrea-dd-dataLte100.npy
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d2dda5dce8674842673a56eac34a77051908f407ead66085b1b84c4d372cba2c
|
| 3 |
+
size 146828479
|
README.md
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# MLSP-Project
|
| 2 |
+
MLSP E9-205 project submission of Vura T N M Srinadh (24923) and Islavath Ganesh (24976)
|
classicml.ipynb
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
graphformers (1).ipynb
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
proj2.ipynb
ADDED
|
@@ -0,0 +1,843 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cells": [
|
| 3 |
+
{
|
| 4 |
+
"cell_type": "code",
|
| 5 |
+
"execution_count": 1,
|
| 6 |
+
"id": "f49d031a",
|
| 7 |
+
"metadata": {},
|
| 8 |
+
"outputs": [],
|
| 9 |
+
"source": [
|
| 10 |
+
"import torch\n",
|
| 11 |
+
"from torch_geometric.data import Data\n",
|
| 12 |
+
"import torch_geometric.transforms as T\n",
|
| 13 |
+
"from torch_geometric.data import InMemoryDataset\n",
|
| 14 |
+
"from torch_geometric.loader import DataLoader\n",
|
| 15 |
+
"import numpy as np\n",
|
| 16 |
+
"import time"
|
| 17 |
+
]
|
| 18 |
+
},
|
| 19 |
+
{
|
| 20 |
+
"cell_type": "code",
|
| 21 |
+
"execution_count": 2,
|
| 22 |
+
"id": "cd1e01d3",
|
| 23 |
+
"metadata": {},
|
| 24 |
+
"outputs": [],
|
| 25 |
+
"source": [
|
| 26 |
+
"path =\"Copy of andrea-dd-dataLte100.npy\"\n",
|
| 27 |
+
"# path=\"/content/drive/MyDrive/Barc Datasets/andrea-juslin-dataGt100.npy\"\n",
|
| 28 |
+
"data = np.load(path, allow_pickle=True)\n",
|
| 29 |
+
"data = data[None][0]\n",
|
| 30 |
+
"dataXorig = []\n",
|
| 31 |
+
"dataY = []\n",
|
| 32 |
+
"dataPe = []\n",
|
| 33 |
+
"for x, y, z in zip(data['x'], data['y'], data['z']):\n",
|
| 34 |
+
" if y > 3: continue # We only want 0 and 1 labels for this activity\n",
|
| 35 |
+
" dataXorig.append(x)\n",
|
| 36 |
+
" dataY.append(y)\n",
|
| 37 |
+
" dataPe.append(z['pe'])"
|
| 38 |
+
]
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"cell_type": "code",
|
| 42 |
+
"execution_count": 3,
|
| 43 |
+
"id": "4f8c43ad",
|
| 44 |
+
"metadata": {},
|
| 45 |
+
"outputs": [],
|
| 46 |
+
"source": [
|
| 47 |
+
"featLen = 30 # we are restricting to first thirty nearest neighbours\n",
|
| 48 |
+
"dataAtoms = torch.tensor(np.array([x[0][:featLen+1] for x in dataXorig]), dtype=torch.float)\n",
|
| 49 |
+
"dataX = torch.tensor(np.array([x[1][:featLen+1] for x in dataXorig]).reshape(len(dataAtoms),featLen+1,1), dtype=torch.float)\n",
|
| 50 |
+
"# dataY = torch.nn.functional.one_hot(torch.tensor(dataY, dtype=torch.long),num_classes=4).float()\n",
|
| 51 |
+
"dataY=torch.tensor(dataY, dtype=torch.long)\n",
|
| 52 |
+
"#dataX = dataAtoms"
|
| 53 |
+
]
|
| 54 |
+
},
|
| 55 |
+
{
|
| 56 |
+
"cell_type": "code",
|
| 57 |
+
"execution_count": 4,
|
| 58 |
+
"id": "70d2090f",
|
| 59 |
+
"metadata": {},
|
| 60 |
+
"outputs": [
|
| 61 |
+
{
|
| 62 |
+
"data": {
|
| 63 |
+
"text/plain": [
|
| 64 |
+
"(torch.Size([42699, 31, 4]),\n",
|
| 65 |
+
" torch.Size([42699, 31, 1]),\n",
|
| 66 |
+
" torch.Size([42699]),\n",
|
| 67 |
+
" torch.Size([42699, 31, 3]))"
|
| 68 |
+
]
|
| 69 |
+
},
|
| 70 |
+
"execution_count": 4,
|
| 71 |
+
"metadata": {},
|
| 72 |
+
"output_type": "execute_result"
|
| 73 |
+
}
|
| 74 |
+
],
|
| 75 |
+
"source": [
|
| 76 |
+
"dataXAll = torch.concat((dataX, dataAtoms), axis=2)\n",
|
| 77 |
+
"dataXAll.shape, dataX.shape, dataY.shape, dataAtoms.shape"
|
| 78 |
+
]
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"cell_type": "code",
|
| 82 |
+
"execution_count": 5,
|
| 83 |
+
"id": "832769fa",
|
| 84 |
+
"metadata": {},
|
| 85 |
+
"outputs": [
|
| 86 |
+
{
|
| 87 |
+
"data": {
|
| 88 |
+
"text/plain": [
|
| 89 |
+
"tensor([0, 0, 0, ..., 1, 1, 1])"
|
| 90 |
+
]
|
| 91 |
+
},
|
| 92 |
+
"execution_count": 5,
|
| 93 |
+
"metadata": {},
|
| 94 |
+
"output_type": "execute_result"
|
| 95 |
+
}
|
| 96 |
+
],
|
| 97 |
+
"source": [
|
| 98 |
+
"dataY"
|
| 99 |
+
]
|
| 100 |
+
},
|
| 101 |
+
{
|
| 102 |
+
"cell_type": "code",
|
| 103 |
+
"execution_count": 6,
|
| 104 |
+
"id": "0d430c0d",
|
| 105 |
+
"metadata": {},
|
| 106 |
+
"outputs": [],
|
| 107 |
+
"source": [
|
| 108 |
+
"posData = []\n",
|
| 109 |
+
"for dt, y in zip(dataAtoms, dataY):\n",
|
| 110 |
+
" data = Data(x= dt, pos=dt, y = y)#, pre_transform=T.RadiusGraph(r=4.0), transform=T.Distance())\n",
|
| 111 |
+
" data.validate(raise_on_error=True)\n",
|
| 112 |
+
" posData.append(data)\n",
|
| 113 |
+
"distData = []\n",
|
| 114 |
+
"for dt, x, y in zip(dataAtoms, dataX, dataY):\n",
|
| 115 |
+
" data = Data(x= x, pos=dt, y = y)#, pre_transform=T.RadiusGraph(r=4.0), transform=T.Distance())\n",
|
| 116 |
+
" data.validate(raise_on_error=True)\n",
|
| 117 |
+
" distData.append(data)\n",
|
| 118 |
+
"allData = posData"
|
| 119 |
+
]
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"cell_type": "code",
|
| 123 |
+
"execution_count": 7,
|
| 124 |
+
"id": "ba1c05ab",
|
| 125 |
+
"metadata": {},
|
| 126 |
+
"outputs": [],
|
| 127 |
+
"source": [
|
| 128 |
+
"dir(data)\n",
|
| 129 |
+
"data.edge_weight"
|
| 130 |
+
]
|
| 131 |
+
},
|
| 132 |
+
{
|
| 133 |
+
"cell_type": "code",
|
| 134 |
+
"execution_count": 8,
|
| 135 |
+
"id": "47afb2ee",
|
| 136 |
+
"metadata": {},
|
| 137 |
+
"outputs": [],
|
| 138 |
+
"source": [
|
| 139 |
+
"import torch\n",
|
| 140 |
+
"from torch_geometric.data import Data\n",
|
| 141 |
+
"import torch_geometric.transforms as T\n",
|
| 142 |
+
"from torch_geometric.data import InMemoryDataset\n",
|
| 143 |
+
"from torch_geometric.loader import DataLoader\n",
|
| 144 |
+
"import numpy as np\n",
|
| 145 |
+
"import time"
|
| 146 |
+
]
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"cell_type": "code",
|
| 150 |
+
"execution_count": 9,
|
| 151 |
+
"id": "80b6c9ac",
|
| 152 |
+
"metadata": {},
|
| 153 |
+
"outputs": [],
|
| 154 |
+
"source": [
|
| 155 |
+
"device=\"cuda\" if torch.cuda.is_available() else \"cpu\""
|
| 156 |
+
]
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"cell_type": "code",
|
| 160 |
+
"execution_count": 10,
|
| 161 |
+
"id": "06225cca",
|
| 162 |
+
"metadata": {},
|
| 163 |
+
"outputs": [],
|
| 164 |
+
"source": [
|
| 165 |
+
"class MyDataset(InMemoryDataset):\n",
|
| 166 |
+
" def __init__(self, root, data_list, transform=None, pre_transform=None, pre_filter=None):\n",
|
| 167 |
+
" self.data_list = data_list\n",
|
| 168 |
+
" super().__init__(root, transform, pre_transform, pre_filter)\n",
|
| 169 |
+
" self.data, self.slices = torch.load(self.processed_paths[0])\n",
|
| 170 |
+
"\n",
|
| 171 |
+
" @property\n",
|
| 172 |
+
" def raw_file_names(self):\n",
|
| 173 |
+
" return ['mydata']\n",
|
| 174 |
+
"\n",
|
| 175 |
+
" @property\n",
|
| 176 |
+
" def processed_file_names(self):\n",
|
| 177 |
+
" return ['data.pt']\n",
|
| 178 |
+
"\n",
|
| 179 |
+
" def download(self):\n",
|
| 180 |
+
" # Download to `self.raw_dir`.\n",
|
| 181 |
+
" pass\n",
|
| 182 |
+
"\n",
|
| 183 |
+
" def process(self):\n",
|
| 184 |
+
" # Read data into huge `Data` list.\n",
|
| 185 |
+
" data_list = self.data_list\n",
|
| 186 |
+
"\n",
|
| 187 |
+
" if self.pre_filter is not None:\n",
|
| 188 |
+
" data_list = [data for data in data_list if self.pre_filter(data)]\n",
|
| 189 |
+
"\n",
|
| 190 |
+
" if self.pre_transform is not None:\n",
|
| 191 |
+
" data_list = [self.pre_transform(data) for data in data_list]\n",
|
| 192 |
+
"\n",
|
| 193 |
+
" data, slices = self.collate(data_list)\n",
|
| 194 |
+
" torch.save((data, slices), self.processed_paths[0])"
|
| 195 |
+
]
|
| 196 |
+
},
|
| 197 |
+
{
|
| 198 |
+
"cell_type": "code",
|
| 199 |
+
"execution_count": 11,
|
| 200 |
+
"id": "91654650",
|
| 201 |
+
"metadata": {},
|
| 202 |
+
"outputs": [
|
| 203 |
+
{
|
| 204 |
+
"name": "stderr",
|
| 205 |
+
"output_type": "stream",
|
| 206 |
+
"text": [
|
| 207 |
+
"Processing...\n",
|
| 208 |
+
"Done!\n",
|
| 209 |
+
"C:\\Users\\SrinadhVura\\AppData\\Local\\Temp\\ipykernel_18728\\1359331624.py:5: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.\n",
|
| 210 |
+
" self.data, self.slices = torch.load(self.processed_paths[0])\n"
|
| 211 |
+
]
|
| 212 |
+
}
|
| 213 |
+
],
|
| 214 |
+
"source": [
|
| 215 |
+
"# !rm -rf ./data/processed/\n",
|
| 216 |
+
"import shutil\n",
|
| 217 |
+
"shutil.rmtree('./data/processed/', ignore_errors=True)\n",
|
| 218 |
+
"dataset = MyDataset(\"./data\",allData, pre_transform=T.Compose([T.RadiusGraph(r=2.0), T.Distance()]))\n",
|
| 219 |
+
"#dataset = MyDataset(\"./data\",allData, pre_transform=T.Compose([T.RadiusGraph(r=2.0), T.Distance()]))"
|
| 220 |
+
]
|
| 221 |
+
},
|
| 222 |
+
{
|
| 223 |
+
"cell_type": "code",
|
| 224 |
+
"execution_count": 15,
|
| 225 |
+
"id": "3edad4e2",
|
| 226 |
+
"metadata": {},
|
| 227 |
+
"outputs": [],
|
| 228 |
+
"source": [
|
| 229 |
+
"from torch_geometric.nn import GATv2Conv\n",
|
| 230 |
+
"import torch.nn.functional as F\n",
|
| 231 |
+
"from torch_geometric.nn.glob import global_mean_pool\n",
|
| 232 |
+
"class GATv2(torch.nn.Module):\n",
|
| 233 |
+
" def __init__(self,dim_h,heads=4):\n",
|
| 234 |
+
" super().__init__()\n",
|
| 235 |
+
" torch.manual_seed(55)\n",
|
| 236 |
+
" edge_dim = dataset[0].edge_attr.shape[1]\n",
|
| 237 |
+
" self.gat1=GATv2Conv(dataset.num_node_features,dim_h,heads=heads,concat=True,edge_dim=edge_dim)\n",
|
| 238 |
+
" self.gat2=GATv2Conv(dim_h*heads,dim_h,heads=heads,concat=True,edge_dim=edge_dim)\n",
|
| 239 |
+
" self.lin=torch.nn.Linear(dim_h*heads,dataset.num_classes)\n",
|
| 240 |
+
" def forward(self,data):\n",
|
| 241 |
+
" x,edge_index,edge_attr,batch=data.x,data.edge_index,data.edge_attr,data.batch\n",
|
| 242 |
+
" x=self.gat1(x,edge_index,edge_attr)\n",
|
| 243 |
+
" x=F.elu(x)\n",
|
| 244 |
+
" x=self.gat2(x,edge_index,edge_attr)\n",
|
| 245 |
+
" x=F.elu(x)\n",
|
| 246 |
+
" x=global_mean_pool(x,batch)\n",
|
| 247 |
+
" x=F.dropout(x,0.3,training=self.training)\n",
|
| 248 |
+
" x=self.lin(x)\n",
|
| 249 |
+
" return x\n"
|
| 250 |
+
]
|
| 251 |
+
},
|
| 252 |
+
{
|
| 253 |
+
"cell_type": "code",
|
| 254 |
+
"execution_count": 16,
|
| 255 |
+
"id": "58bb1ac9",
|
| 256 |
+
"metadata": {},
|
| 257 |
+
"outputs": [],
|
| 258 |
+
"source": [
|
| 259 |
+
"dataset = dataset.shuffle()\n",
|
| 260 |
+
"train_size = int(0.8 * len(dataset))\n",
|
| 261 |
+
"test_size = len(dataset) - train_size\n",
|
| 262 |
+
"train_data,test_data=dataset[:train_size],dataset[train_size:]\n",
|
| 263 |
+
"train_loader = DataLoader(train_data, batch_size=32, shuffle=True)\n",
|
| 264 |
+
"test_loader = DataLoader(test_data, batch_size=32, shuffle=False)"
|
| 265 |
+
]
|
| 266 |
+
},
|
| 267 |
+
{
|
| 268 |
+
"cell_type": "code",
|
| 269 |
+
"execution_count": 17,
|
| 270 |
+
"id": "2d6479c6",
|
| 271 |
+
"metadata": {},
|
| 272 |
+
"outputs": [
|
| 273 |
+
{
|
| 274 |
+
"name": "stdout",
|
| 275 |
+
"output_type": "stream",
|
| 276 |
+
"text": [
|
| 277 |
+
"Epoch: 001, Train Acc: 0.4957, Test Acc: 0.4900, Time: 25.8774 seconds\n",
|
| 278 |
+
"Epoch: 002, Train Acc: 0.6084, Test Acc: 0.6093, Time: 24.4807 seconds\n",
|
| 279 |
+
"Epoch: 003, Train Acc: 0.6459, Test Acc: 0.6454, Time: 24.6017 seconds\n",
|
| 280 |
+
"Epoch: 004, Train Acc: 0.6813, Test Acc: 0.6749, Time: 24.6508 seconds\n",
|
| 281 |
+
"Epoch: 005, Train Acc: 0.6887, Test Acc: 0.6850, Time: 24.8438 seconds\n",
|
| 282 |
+
"Epoch: 006, Train Acc: 0.6733, Test Acc: 0.6679, Time: 24.9810 seconds\n",
|
| 283 |
+
"Epoch: 007, Train Acc: 0.7105, Test Acc: 0.7028, Time: 25.4978 seconds\n",
|
| 284 |
+
"Epoch: 008, Train Acc: 0.7156, Test Acc: 0.7070, Time: 26.5686 seconds\n",
|
| 285 |
+
"Epoch: 009, Train Acc: 0.7192, Test Acc: 0.7057, Time: 25.5068 seconds\n",
|
| 286 |
+
"Epoch: 010, Train Acc: 0.7452, Test Acc: 0.7364, Time: 24.8775 seconds\n",
|
| 287 |
+
"Epoch: 011, Train Acc: 0.7473, Test Acc: 0.7358, Time: 24.8751 seconds\n",
|
| 288 |
+
"Epoch: 012, Train Acc: 0.7590, Test Acc: 0.7523, Time: 24.8378 seconds\n",
|
| 289 |
+
"Epoch: 013, Train Acc: 0.7534, Test Acc: 0.7441, Time: 24.8861 seconds\n",
|
| 290 |
+
"Epoch: 014, Train Acc: 0.7390, Test Acc: 0.7297, Time: 24.8664 seconds\n",
|
| 291 |
+
"Epoch: 015, Train Acc: 0.7719, Test Acc: 0.7642, Time: 24.7913 seconds\n",
|
| 292 |
+
"Epoch: 016, Train Acc: 0.7905, Test Acc: 0.7778, Time: 24.7148 seconds\n",
|
| 293 |
+
"Epoch: 017, Train Acc: 0.7631, Test Acc: 0.7585, Time: 24.8516 seconds\n",
|
| 294 |
+
"Epoch: 018, Train Acc: 0.7694, Test Acc: 0.7562, Time: 24.8459 seconds\n",
|
| 295 |
+
"Epoch: 019, Train Acc: 0.7915, Test Acc: 0.7809, Time: 25.7345 seconds\n",
|
| 296 |
+
"Epoch: 020, Train Acc: 0.7866, Test Acc: 0.7721, Time: 25.8023 seconds\n",
|
| 297 |
+
"Epoch: 021, Train Acc: 0.7834, Test Acc: 0.7715, Time: 26.0532 seconds\n",
|
| 298 |
+
"Epoch: 022, Train Acc: 0.7917, Test Acc: 0.7799, Time: 25.6241 seconds\n",
|
| 299 |
+
"Epoch: 023, Train Acc: 0.7874, Test Acc: 0.7795, Time: 25.5449 seconds\n",
|
| 300 |
+
"Epoch: 024, Train Acc: 0.8107, Test Acc: 0.7905, Time: 25.6376 seconds\n",
|
| 301 |
+
"Epoch: 025, Train Acc: 0.8006, Test Acc: 0.7877, Time: 25.4805 seconds\n",
|
| 302 |
+
"Epoch: 026, Train Acc: 0.8106, Test Acc: 0.7940, Time: 25.5074 seconds\n",
|
| 303 |
+
"Epoch: 027, Train Acc: 0.8090, Test Acc: 0.7974, Time: 25.8159 seconds\n",
|
| 304 |
+
"Epoch: 028, Train Acc: 0.7693, Test Acc: 0.7540, Time: 25.4768 seconds\n",
|
| 305 |
+
"Epoch: 029, Train Acc: 0.8152, Test Acc: 0.7982, Time: 25.3734 seconds\n",
|
| 306 |
+
"Epoch: 030, Train Acc: 0.8120, Test Acc: 0.7951, Time: 25.2074 seconds\n",
|
| 307 |
+
"Epoch: 031, Train Acc: 0.7903, Test Acc: 0.7717, Time: 25.2981 seconds\n",
|
| 308 |
+
"Epoch: 032, Train Acc: 0.7993, Test Acc: 0.7815, Time: 25.2922 seconds\n",
|
| 309 |
+
"Epoch: 033, Train Acc: 0.8092, Test Acc: 0.7857, Time: 25.2716 seconds\n",
|
| 310 |
+
"Epoch: 034, Train Acc: 0.8086, Test Acc: 0.7821, Time: 25.1796 seconds\n",
|
| 311 |
+
"Epoch: 035, Train Acc: 0.8158, Test Acc: 0.7929, Time: 25.2028 seconds\n",
|
| 312 |
+
"Epoch: 036, Train Acc: 0.8133, Test Acc: 0.7926, Time: 25.1788 seconds\n",
|
| 313 |
+
"Epoch: 037, Train Acc: 0.8030, Test Acc: 0.7852, Time: 25.1870 seconds\n",
|
| 314 |
+
"Epoch: 038, Train Acc: 0.8300, Test Acc: 0.8062, Time: 25.2557 seconds\n",
|
| 315 |
+
"Epoch: 039, Train Acc: 0.8002, Test Acc: 0.7762, Time: 25.2280 seconds\n",
|
| 316 |
+
"Epoch: 040, Train Acc: 0.8199, Test Acc: 0.7973, Time: 25.2002 seconds\n",
|
| 317 |
+
"Epoch: 041, Train Acc: 0.8112, Test Acc: 0.7879, Time: 25.1714 seconds\n",
|
| 318 |
+
"Epoch: 042, Train Acc: 0.8256, Test Acc: 0.8116, Time: 25.2478 seconds\n",
|
| 319 |
+
"Epoch: 043, Train Acc: 0.8447, Test Acc: 0.8215, Time: 25.3352 seconds\n",
|
| 320 |
+
"Epoch: 044, Train Acc: 0.8315, Test Acc: 0.8033, Time: 26.1272 seconds\n",
|
| 321 |
+
"Epoch: 045, Train Acc: 0.8426, Test Acc: 0.8194, Time: 25.4368 seconds\n",
|
| 322 |
+
"Epoch: 046, Train Acc: 0.8236, Test Acc: 0.7975, Time: 25.2172 seconds\n",
|
| 323 |
+
"Epoch: 047, Train Acc: 0.8340, Test Acc: 0.8093, Time: 25.4511 seconds\n",
|
| 324 |
+
"Epoch: 048, Train Acc: 0.8483, Test Acc: 0.8281, Time: 25.2266 seconds\n",
|
| 325 |
+
"Epoch: 049, Train Acc: 0.8475, Test Acc: 0.8275, Time: 25.1697 seconds\n",
|
| 326 |
+
"TOtal time taken:1238.5092 seconds\n"
|
| 327 |
+
]
|
| 328 |
+
}
|
| 329 |
+
],
|
| 330 |
+
"source": [
|
| 331 |
+
"\n",
|
| 332 |
+
"model = GATv2(128).to(device)\n",
|
| 333 |
+
"optimizer = torch.optim.Adam(model.parameters(), lr=0.001)\n",
|
| 334 |
+
"criterion = torch.nn.CrossEntropyLoss()\n",
|
| 335 |
+
"\n",
|
| 336 |
+
"def train():\n",
|
| 337 |
+
" model.train()\n",
|
| 338 |
+
"\n",
|
| 339 |
+
" for data in train_loader: # Iterate in batches over the training dataset.\n",
|
| 340 |
+
" out = model(data.to(device)) # Perform a single forward pass.\n",
|
| 341 |
+
" loss = criterion(out, data.y) # Compute the loss.\n",
|
| 342 |
+
" loss.backward() # Derive gradients.\n",
|
| 343 |
+
" optimizer.step() # Update parameters based on gradients.\n",
|
| 344 |
+
" optimizer.zero_grad() # Clear gradients.\n",
|
| 345 |
+
"\n",
|
| 346 |
+
"def test(loader):\n",
|
| 347 |
+
" model.eval()\n",
|
| 348 |
+
"\n",
|
| 349 |
+
" correct = 0\n",
|
| 350 |
+
" for data in loader: # Iterate in batches over the training/test dataset.\n",
|
| 351 |
+
" out = model(data.to(device))\n",
|
| 352 |
+
" pred = out.argmax(dim=1) # Use the class with highest probability.\n",
|
| 353 |
+
" correct += int((pred == data.y).sum()) # Check against ground-truth labels.\n",
|
| 354 |
+
" return correct / len(loader.dataset) # Derive ratio of correct predictions.\n",
|
| 355 |
+
"\n",
|
| 356 |
+
"\n",
|
| 357 |
+
"acc,t=[],[]\n",
|
| 358 |
+
"a=time.time()\n",
|
| 359 |
+
"for epoch in range(1, 50):\n",
|
| 360 |
+
" x=time.time()\n",
|
| 361 |
+
" train()\n",
|
| 362 |
+
" train_acc = test(train_loader)\n",
|
| 363 |
+
" test_acc = test(test_loader)\n",
|
| 364 |
+
" y=time.time()\n",
|
| 365 |
+
" acc.append(test_acc)\n",
|
| 366 |
+
" t.append(y-x)\n",
|
| 367 |
+
" print(f'Epoch: {epoch:03d}, Train Acc: {train_acc:.4f}, Test Acc: {test_acc:.4f}, Time: {(y-x):.4f} seconds')\n",
|
| 368 |
+
"b=time.time()\n",
|
| 369 |
+
"print(f'TOtal time taken:{(b-a):.4f} seconds')"
|
| 370 |
+
]
|
| 371 |
+
},
|
| 372 |
+
{
|
| 373 |
+
"cell_type": "code",
|
| 374 |
+
"execution_count": 18,
|
| 375 |
+
"id": "2d9dbd7c",
|
| 376 |
+
"metadata": {},
|
| 377 |
+
"outputs": [
|
| 378 |
+
{
|
| 379 |
+
"name": "stdout",
|
| 380 |
+
"output_type": "stream",
|
| 381 |
+
"text": [
|
| 382 |
+
"Epoch: 001, Train Acc: 0.8265, Test Acc: 0.7945, Time: 24.9625 seconds\n",
|
| 383 |
+
"Epoch: 002, Train Acc: 0.8503, Test Acc: 0.8258, Time: 24.9976 seconds\n",
|
| 384 |
+
"Epoch: 003, Train Acc: 0.8506, Test Acc: 0.8281, Time: 25.1676 seconds\n",
|
| 385 |
+
"Epoch: 004, Train Acc: 0.8372, Test Acc: 0.8184, Time: 25.4236 seconds\n",
|
| 386 |
+
"Epoch: 005, Train Acc: 0.8557, Test Acc: 0.8248, Time: 27.4433 seconds\n",
|
| 387 |
+
"Epoch: 006, Train Acc: 0.8488, Test Acc: 0.8201, Time: 26.2164 seconds\n",
|
| 388 |
+
"Epoch: 007, Train Acc: 0.8493, Test Acc: 0.8219, Time: 28.7789 seconds\n",
|
| 389 |
+
"Epoch: 008, Train Acc: 0.8429, Test Acc: 0.8165, Time: 25.7886 seconds\n",
|
| 390 |
+
"Epoch: 009, Train Acc: 0.8613, Test Acc: 0.8276, Time: 26.8074 seconds\n",
|
| 391 |
+
"Epoch: 010, Train Acc: 0.8528, Test Acc: 0.8196, Time: 26.9662 seconds\n",
|
| 392 |
+
"Epoch: 011, Train Acc: 0.8372, Test Acc: 0.8053, Time: 25.6793 seconds\n",
|
| 393 |
+
"Epoch: 012, Train Acc: 0.8352, Test Acc: 0.8151, Time: 25.4075 seconds\n",
|
| 394 |
+
"Epoch: 013, Train Acc: 0.8603, Test Acc: 0.8337, Time: 25.2606 seconds\n",
|
| 395 |
+
"Epoch: 014, Train Acc: 0.8521, Test Acc: 0.8254, Time: 25.2944 seconds\n",
|
| 396 |
+
"Epoch: 015, Train Acc: 0.8498, Test Acc: 0.8143, Time: 25.2541 seconds\n",
|
| 397 |
+
"Epoch: 016, Train Acc: 0.8547, Test Acc: 0.8205, Time: 25.2576 seconds\n",
|
| 398 |
+
"Epoch: 017, Train Acc: 0.8482, Test Acc: 0.8220, Time: 25.1988 seconds\n",
|
| 399 |
+
"Epoch: 018, Train Acc: 0.8525, Test Acc: 0.8296, Time: 25.2256 seconds\n",
|
| 400 |
+
"Epoch: 019, Train Acc: 0.8595, Test Acc: 0.8259, Time: 25.4156 seconds\n",
|
| 401 |
+
"Epoch: 020, Train Acc: 0.8589, Test Acc: 0.8261, Time: 25.2268 seconds\n",
|
| 402 |
+
"Epoch: 021, Train Acc: 0.8658, Test Acc: 0.8304, Time: 25.4441 seconds\n",
|
| 403 |
+
"Epoch: 022, Train Acc: 0.8352, Test Acc: 0.8070, Time: 25.5956 seconds\n",
|
| 404 |
+
"Epoch: 023, Train Acc: 0.8724, Test Acc: 0.8299, Time: 25.4439 seconds\n",
|
| 405 |
+
"Epoch: 024, Train Acc: 0.8242, Test Acc: 0.7918, Time: 25.2510 seconds\n",
|
| 406 |
+
"Epoch: 025, Train Acc: 0.8627, Test Acc: 0.8307, Time: 26.1188 seconds\n",
|
| 407 |
+
"Epoch: 026, Train Acc: 0.8717, Test Acc: 0.8348, Time: 26.0500 seconds\n",
|
| 408 |
+
"Epoch: 027, Train Acc: 0.8544, Test Acc: 0.8184, Time: 25.5288 seconds\n",
|
| 409 |
+
"Epoch: 028, Train Acc: 0.7814, Test Acc: 0.7487, Time: 25.7745 seconds\n",
|
| 410 |
+
"Epoch: 029, Train Acc: 0.8569, Test Acc: 0.8292, Time: 25.3209 seconds\n",
|
| 411 |
+
"Epoch: 030, Train Acc: 0.8623, Test Acc: 0.8292, Time: 25.2030 seconds\n",
|
| 412 |
+
"Epoch: 031, Train Acc: 0.8640, Test Acc: 0.8287, Time: 25.9134 seconds\n",
|
| 413 |
+
"Epoch: 032, Train Acc: 0.8660, Test Acc: 0.8343, Time: 25.8696 seconds\n",
|
| 414 |
+
"Epoch: 033, Train Acc: 0.8541, Test Acc: 0.8167, Time: 25.1739 seconds\n",
|
| 415 |
+
"Epoch: 034, Train Acc: 0.8516, Test Acc: 0.8204, Time: 25.2224 seconds\n",
|
| 416 |
+
"Epoch: 035, Train Acc: 0.8651, Test Acc: 0.8274, Time: 25.4136 seconds\n",
|
| 417 |
+
"Epoch: 036, Train Acc: 0.8691, Test Acc: 0.8329, Time: 25.2840 seconds\n",
|
| 418 |
+
"Epoch: 037, Train Acc: 0.8794, Test Acc: 0.8351, Time: 25.4689 seconds\n",
|
| 419 |
+
"Epoch: 038, Train Acc: 0.8390, Test Acc: 0.8005, Time: 25.4441 seconds\n",
|
| 420 |
+
"Epoch: 039, Train Acc: 0.8515, Test Acc: 0.8180, Time: 25.4249 seconds\n",
|
| 421 |
+
"Epoch: 040, Train Acc: 0.8535, Test Acc: 0.8221, Time: 25.4681 seconds\n",
|
| 422 |
+
"Epoch: 041, Train Acc: 0.8727, Test Acc: 0.8391, Time: 25.1276 seconds\n",
|
| 423 |
+
"Epoch: 042, Train Acc: 0.8796, Test Acc: 0.8400, Time: 24.9614 seconds\n",
|
| 424 |
+
"Epoch: 043, Train Acc: 0.8708, Test Acc: 0.8330, Time: 25.3893 seconds\n",
|
| 425 |
+
"Epoch: 044, Train Acc: 0.8717, Test Acc: 0.8348, Time: 25.6369 seconds\n",
|
| 426 |
+
"Epoch: 045, Train Acc: 0.8654, Test Acc: 0.8273, Time: 25.1120 seconds\n",
|
| 427 |
+
"Epoch: 046, Train Acc: 0.8614, Test Acc: 0.8251, Time: 25.0642 seconds\n",
|
| 428 |
+
"Epoch: 047, Train Acc: 0.8609, Test Acc: 0.8207, Time: 25.0014 seconds\n",
|
| 429 |
+
"Epoch: 048, Train Acc: 0.8761, Test Acc: 0.8411, Time: 25.0245 seconds\n",
|
| 430 |
+
"Epoch: 049, Train Acc: 0.8633, Test Acc: 0.8239, Time: 25.0614 seconds\n",
|
| 431 |
+
"TOtal time taken:2581.8546 seconds\n"
|
| 432 |
+
]
|
| 433 |
+
}
|
| 434 |
+
],
|
| 435 |
+
"source": [
|
| 436 |
+
"for epoch in range(1, 50):\n",
|
| 437 |
+
" x=time.time()\n",
|
| 438 |
+
" train()\n",
|
| 439 |
+
" train_acc = test(train_loader)\n",
|
| 440 |
+
" test_acc = test(test_loader)\n",
|
| 441 |
+
" y=time.time()\n",
|
| 442 |
+
" acc.append(test_acc)\n",
|
| 443 |
+
" t.append(y-x)\n",
|
| 444 |
+
" print(f'Epoch: {epoch:03d}, Train Acc: {train_acc:.4f}, Test Acc: {test_acc:.4f}, Time: {(y-x):.4f} seconds')\n",
|
| 445 |
+
"b=time.time()\n",
|
| 446 |
+
"print(f'TOtal time taken:{(b-a):.4f} seconds')"
|
| 447 |
+
]
|
| 448 |
+
},
|
| 449 |
+
{
|
| 450 |
+
"cell_type": "code",
|
| 451 |
+
"execution_count": 19,
|
| 452 |
+
"id": "d7c43502",
|
| 453 |
+
"metadata": {},
|
| 454 |
+
"outputs": [
|
| 455 |
+
{
|
| 456 |
+
"name": "stdout",
|
| 457 |
+
"output_type": "stream",
|
| 458 |
+
"text": [
|
| 459 |
+
"Epoch: 001, Train Acc: 0.8673, Test Acc: 0.8311, Time: 25.3796 seconds\n",
|
| 460 |
+
"Epoch: 002, Train Acc: 0.8462, Test Acc: 0.8125, Time: 25.1563 seconds\n",
|
| 461 |
+
"Epoch: 003, Train Acc: 0.8498, Test Acc: 0.8034, Time: 25.2885 seconds\n",
|
| 462 |
+
"Epoch: 004, Train Acc: 0.8368, Test Acc: 0.8027, Time: 25.2527 seconds\n",
|
| 463 |
+
"Epoch: 005, Train Acc: 0.8805, Test Acc: 0.8356, Time: 26.5682 seconds\n",
|
| 464 |
+
"Epoch: 006, Train Acc: 0.8526, Test Acc: 0.8156, Time: 25.8687 seconds\n",
|
| 465 |
+
"Epoch: 007, Train Acc: 0.8837, Test Acc: 0.8427, Time: 25.6244 seconds\n",
|
| 466 |
+
"Epoch: 008, Train Acc: 0.8823, Test Acc: 0.8358, Time: 25.2722 seconds\n",
|
| 467 |
+
"Epoch: 009, Train Acc: 0.8740, Test Acc: 0.8337, Time: 25.2649 seconds\n",
|
| 468 |
+
"Epoch: 010, Train Acc: 0.8775, Test Acc: 0.8349, Time: 25.5523 seconds\n",
|
| 469 |
+
"Epoch: 011, Train Acc: 0.8671, Test Acc: 0.8299, Time: 25.1311 seconds\n",
|
| 470 |
+
"Epoch: 012, Train Acc: 0.8681, Test Acc: 0.8315, Time: 25.1274 seconds\n",
|
| 471 |
+
"Epoch: 013, Train Acc: 0.8867, Test Acc: 0.8430, Time: 25.1428 seconds\n",
|
| 472 |
+
"Epoch: 014, Train Acc: 0.8808, Test Acc: 0.8370, Time: 25.3863 seconds\n",
|
| 473 |
+
"Epoch: 015, Train Acc: 0.8705, Test Acc: 0.8361, Time: 25.3820 seconds\n",
|
| 474 |
+
"Epoch: 016, Train Acc: 0.8663, Test Acc: 0.8262, Time: 25.3648 seconds\n",
|
| 475 |
+
"Epoch: 017, Train Acc: 0.8764, Test Acc: 0.8315, Time: 25.1957 seconds\n",
|
| 476 |
+
"Epoch: 018, Train Acc: 0.8839, Test Acc: 0.8341, Time: 25.1296 seconds\n",
|
| 477 |
+
"Epoch: 019, Train Acc: 0.8744, Test Acc: 0.8259, Time: 25.1842 seconds\n",
|
| 478 |
+
"Epoch: 020, Train Acc: 0.8728, Test Acc: 0.8272, Time: 25.1960 seconds\n",
|
| 479 |
+
"Epoch: 021, Train Acc: 0.8814, Test Acc: 0.8411, Time: 25.2335 seconds\n",
|
| 480 |
+
"Epoch: 022, Train Acc: 0.8612, Test Acc: 0.8265, Time: 25.2171 seconds\n",
|
| 481 |
+
"Epoch: 023, Train Acc: 0.8821, Test Acc: 0.8402, Time: 25.2572 seconds\n",
|
| 482 |
+
"Epoch: 024, Train Acc: 0.8830, Test Acc: 0.8370, Time: 26.0306 seconds\n",
|
| 483 |
+
"Epoch: 025, Train Acc: 0.8693, Test Acc: 0.8261, Time: 25.1953 seconds\n",
|
| 484 |
+
"Epoch: 026, Train Acc: 0.8698, Test Acc: 0.8288, Time: 25.1328 seconds\n",
|
| 485 |
+
"Epoch: 027, Train Acc: 0.8853, Test Acc: 0.8447, Time: 25.1489 seconds\n",
|
| 486 |
+
"Epoch: 028, Train Acc: 0.8785, Test Acc: 0.8285, Time: 25.2630 seconds\n",
|
| 487 |
+
"Epoch: 029, Train Acc: 0.8882, Test Acc: 0.8385, Time: 25.1945 seconds\n",
|
| 488 |
+
"Epoch: 030, Train Acc: 0.8726, Test Acc: 0.8288, Time: 25.1916 seconds\n",
|
| 489 |
+
"Epoch: 031, Train Acc: 0.8895, Test Acc: 0.8386, Time: 25.2485 seconds\n",
|
| 490 |
+
"Epoch: 032, Train Acc: 0.8636, Test Acc: 0.8295, Time: 25.1767 seconds\n",
|
| 491 |
+
"Epoch: 033, Train Acc: 0.8746, Test Acc: 0.8299, Time: 25.2032 seconds\n",
|
| 492 |
+
"Epoch: 034, Train Acc: 0.8880, Test Acc: 0.8383, Time: 25.5541 seconds\n",
|
| 493 |
+
"Epoch: 035, Train Acc: 0.8822, Test Acc: 0.8385, Time: 25.2495 seconds\n",
|
| 494 |
+
"Epoch: 036, Train Acc: 0.8909, Test Acc: 0.8467, Time: 25.4243 seconds\n",
|
| 495 |
+
"Epoch: 037, Train Acc: 0.8839, Test Acc: 0.8351, Time: 25.5948 seconds\n",
|
| 496 |
+
"Epoch: 038, Train Acc: 0.8821, Test Acc: 0.8341, Time: 25.2591 seconds\n",
|
| 497 |
+
"Epoch: 039, Train Acc: 0.8889, Test Acc: 0.8354, Time: 25.2509 seconds\n",
|
| 498 |
+
"Epoch: 040, Train Acc: 0.8842, Test Acc: 0.8446, Time: 25.2162 seconds\n",
|
| 499 |
+
"Epoch: 041, Train Acc: 0.8636, Test Acc: 0.8126, Time: 25.2191 seconds\n",
|
| 500 |
+
"Epoch: 042, Train Acc: 0.8817, Test Acc: 0.8363, Time: 25.2106 seconds\n",
|
| 501 |
+
"Epoch: 043, Train Acc: 0.8897, Test Acc: 0.8368, Time: 25.2143 seconds\n",
|
| 502 |
+
"Epoch: 044, Train Acc: 0.8777, Test Acc: 0.8316, Time: 25.1635 seconds\n",
|
| 503 |
+
"Epoch: 045, Train Acc: 0.8879, Test Acc: 0.8396, Time: 25.2399 seconds\n",
|
| 504 |
+
"Epoch: 046, Train Acc: 0.8850, Test Acc: 0.8309, Time: 25.1394 seconds\n",
|
| 505 |
+
"Epoch: 047, Train Acc: 0.8801, Test Acc: 0.8281, Time: 25.3154 seconds\n",
|
| 506 |
+
"Epoch: 048, Train Acc: 0.8905, Test Acc: 0.8363, Time: 25.1571 seconds\n",
|
| 507 |
+
"Epoch: 049, Train Acc: 0.8916, Test Acc: 0.8450, Time: 25.1891 seconds\n",
|
| 508 |
+
"TOtal time taken:4324.3672 seconds\n"
|
| 509 |
+
]
|
| 510 |
+
}
|
| 511 |
+
],
|
| 512 |
+
"source": [
|
| 513 |
+
"for epoch in range(1, 50):\n",
|
| 514 |
+
" x=time.time()\n",
|
| 515 |
+
" train()\n",
|
| 516 |
+
" train_acc = test(train_loader)\n",
|
| 517 |
+
" test_acc = test(test_loader)\n",
|
| 518 |
+
" y=time.time()\n",
|
| 519 |
+
" acc.append(test_acc)\n",
|
| 520 |
+
" t.append(y-x)\n",
|
| 521 |
+
" print(f'Epoch: {epoch:03d}, Train Acc: {train_acc:.4f}, Test Acc: {test_acc:.4f}, Time: {(y-x):.4f} seconds')\n",
|
| 522 |
+
"b=time.time()\n",
|
| 523 |
+
"print(f'TOtal time taken:{(b-a):.4f} seconds')"
|
| 524 |
+
]
|
| 525 |
+
},
|
| 526 |
+
{
|
| 527 |
+
"cell_type": "code",
|
| 528 |
+
"execution_count": 20,
|
| 529 |
+
"id": "ccdc2139",
|
| 530 |
+
"metadata": {},
|
| 531 |
+
"outputs": [],
|
| 532 |
+
"source": [
|
| 533 |
+
"import torch\n",
|
| 534 |
+
"import torch.nn as nn\n",
|
| 535 |
+
"import torch.nn.functional as F\n",
|
| 536 |
+
"from torch_geometric.nn import TransformerConv, global_mean_pool\n",
|
| 537 |
+
"\n",
|
| 538 |
+
"class GraphTransformerNet(nn.Module):\n",
|
| 539 |
+
" def __init__(self, in_channels, hidden_channels, out_channels, heads=4, num_layers=3, dropout=0.2):\n",
|
| 540 |
+
" super().__init__()\n",
|
| 541 |
+
" self.convs = nn.ModuleList()\n",
|
| 542 |
+
" self.convs.append(TransformerConv(in_channels, hidden_channels, heads=heads, dropout=dropout))\n",
|
| 543 |
+
" for _ in range(num_layers - 2):\n",
|
| 544 |
+
" self.convs.append(TransformerConv(hidden_channels * heads, hidden_channels, heads=heads, dropout=dropout))\n",
|
| 545 |
+
" self.convs.append(TransformerConv(hidden_channels * heads, hidden_channels, heads=1, dropout=dropout))\n",
|
| 546 |
+
" self.lin = nn.Linear(hidden_channels, out_channels)\n",
|
| 547 |
+
" self.dropout = dropout\n",
|
| 548 |
+
"\n",
|
| 549 |
+
" def forward(self, data):\n",
|
| 550 |
+
" x, edge_index, batch = data.x, data.edge_index, data.batch\n",
|
| 551 |
+
" for conv in self.convs:\n",
|
| 552 |
+
" x = conv(x, edge_index)\n",
|
| 553 |
+
" x = F.relu(x)\n",
|
| 554 |
+
" x = F.dropout(x, p=self.dropout, training=self.training)\n",
|
| 555 |
+
" x = global_mean_pool(x, batch)\n",
|
| 556 |
+
" x = self.lin(x)\n",
|
| 557 |
+
" return x"
|
| 558 |
+
]
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"cell_type": "code",
|
| 562 |
+
"execution_count": 26,
|
| 563 |
+
"id": "66fc81fa",
|
| 564 |
+
"metadata": {},
|
| 565 |
+
"outputs": [
|
| 566 |
+
{
|
| 567 |
+
"name": "stdout",
|
| 568 |
+
"output_type": "stream",
|
| 569 |
+
"text": [
|
| 570 |
+
"Epoch: 001, Train Acc: 0.4981, Test Acc: 0.5008, Time: 21.0400 seconds\n",
|
| 571 |
+
"Epoch: 002, Train Acc: 0.6214, Test Acc: 0.6283, Time: 21.0401 seconds\n",
|
| 572 |
+
"Epoch: 003, Train Acc: 0.6922, Test Acc: 0.6972, Time: 21.3110 seconds\n",
|
| 573 |
+
"Epoch: 004, Train Acc: 0.6340, Test Acc: 0.6347, Time: 21.4084 seconds\n",
|
| 574 |
+
"Epoch: 005, Train Acc: 0.7140, Test Acc: 0.7116, Time: 21.5427 seconds\n",
|
| 575 |
+
"Epoch: 006, Train Acc: 0.7017, Test Acc: 0.6966, Time: 21.5673 seconds\n",
|
| 576 |
+
"Epoch: 007, Train Acc: 0.7422, Test Acc: 0.7420, Time: 21.4987 seconds\n",
|
| 577 |
+
"Epoch: 008, Train Acc: 0.7296, Test Acc: 0.7319, Time: 21.3965 seconds\n",
|
| 578 |
+
"Epoch: 009, Train Acc: 0.7007, Test Acc: 0.7085, Time: 21.3689 seconds\n",
|
| 579 |
+
"Epoch: 010, Train Acc: 0.7653, Test Acc: 0.7651, Time: 21.3604 seconds\n",
|
| 580 |
+
"Epoch: 011, Train Acc: 0.7580, Test Acc: 0.7622, Time: 21.3570 seconds\n",
|
| 581 |
+
"Epoch: 012, Train Acc: 0.7195, Test Acc: 0.7173, Time: 21.3565 seconds\n",
|
| 582 |
+
"Epoch: 013, Train Acc: 0.7629, Test Acc: 0.7655, Time: 21.4036 seconds\n",
|
| 583 |
+
"Epoch: 014, Train Acc: 0.7690, Test Acc: 0.7669, Time: 21.3677 seconds\n",
|
| 584 |
+
"Epoch: 015, Train Acc: 0.7581, Test Acc: 0.7540, Time: 21.4620 seconds\n",
|
| 585 |
+
"Epoch: 016, Train Acc: 0.7948, Test Acc: 0.7952, Time: 21.4100 seconds\n",
|
| 586 |
+
"Epoch: 017, Train Acc: 0.7715, Test Acc: 0.7711, Time: 21.5848 seconds\n",
|
| 587 |
+
"Epoch: 018, Train Acc: 0.7848, Test Acc: 0.7813, Time: 21.4893 seconds\n",
|
| 588 |
+
"Epoch: 019, Train Acc: 0.8042, Test Acc: 0.7989, Time: 21.5196 seconds\n",
|
| 589 |
+
"Epoch: 020, Train Acc: 0.7848, Test Acc: 0.7852, Time: 21.4080 seconds\n",
|
| 590 |
+
"Epoch: 021, Train Acc: 0.7854, Test Acc: 0.7845, Time: 21.3282 seconds\n",
|
| 591 |
+
"Epoch: 022, Train Acc: 0.7995, Test Acc: 0.7979, Time: 21.3296 seconds\n",
|
| 592 |
+
"Epoch: 023, Train Acc: 0.8046, Test Acc: 0.8013, Time: 21.2942 seconds\n",
|
| 593 |
+
"Epoch: 024, Train Acc: 0.7934, Test Acc: 0.7945, Time: 21.2278 seconds\n",
|
| 594 |
+
"Epoch: 025, Train Acc: 0.7872, Test Acc: 0.7893, Time: 21.3140 seconds\n",
|
| 595 |
+
"Epoch: 026, Train Acc: 0.8161, Test Acc: 0.8152, Time: 21.2853 seconds\n",
|
| 596 |
+
"Epoch: 027, Train Acc: 0.8057, Test Acc: 0.8043, Time: 21.2965 seconds\n",
|
| 597 |
+
"Epoch: 028, Train Acc: 0.8098, Test Acc: 0.8067, Time: 21.2929 seconds\n",
|
| 598 |
+
"Epoch: 029, Train Acc: 0.8096, Test Acc: 0.8034, Time: 21.3359 seconds\n",
|
| 599 |
+
"TOtal time taken:619.6132 seconds\n"
|
| 600 |
+
]
|
| 601 |
+
}
|
| 602 |
+
],
|
| 603 |
+
"source": [
|
| 604 |
+
"in_channels = dataset.num_node_features\n",
|
| 605 |
+
"hidden_channels = 128\n",
|
| 606 |
+
"out_channels = dataset.num_classes\n",
|
| 607 |
+
"model = GraphTransformerNet(in_channels, hidden_channels, out_channels).to(device)\n",
|
| 608 |
+
"optimizer = torch.optim.Adam(model.parameters(), lr=0.001)\n",
|
| 609 |
+
"criterion = torch.nn.CrossEntropyLoss()\n",
|
| 610 |
+
"\n",
|
| 611 |
+
"def train():\n",
|
| 612 |
+
" model.train()\n",
|
| 613 |
+
"\n",
|
| 614 |
+
" for data in train_loader: # Iterate in batches over the training dataset.\n",
|
| 615 |
+
" out = model(data.to(device)) # Perform a single forward pass.\n",
|
| 616 |
+
" loss = criterion(out, data.y) # Compute the loss.\n",
|
| 617 |
+
" loss.backward() # Derive gradients.\n",
|
| 618 |
+
" optimizer.step() # Update parameters based on gradients.\n",
|
| 619 |
+
" optimizer.zero_grad() # Clear gradients.\n",
|
| 620 |
+
"\n",
|
| 621 |
+
"def test(loader):\n",
|
| 622 |
+
" model.eval()\n",
|
| 623 |
+
"\n",
|
| 624 |
+
" correct = 0\n",
|
| 625 |
+
" for data in loader: # Iterate in batches over the training/test dataset.\n",
|
| 626 |
+
" out = model(data.to(device))\n",
|
| 627 |
+
" pred = out.argmax(dim=1) # Use the class with highest probability.\n",
|
| 628 |
+
" correct += int((pred == data.y).sum()) # Check against ground-truth labels.\n",
|
| 629 |
+
" return correct / len(loader.dataset) # Derive ratio of correct predictions.\n",
|
| 630 |
+
"\n",
|
| 631 |
+
"\n",
|
| 632 |
+
"acc,t=[],[]\n",
|
| 633 |
+
"a=time.time()\n",
|
| 634 |
+
"for epoch in range(1, 30):\n",
|
| 635 |
+
" x=time.time()\n",
|
| 636 |
+
" train()\n",
|
| 637 |
+
" train_acc = test(train_loader)\n",
|
| 638 |
+
" test_acc = test(test_loader)\n",
|
| 639 |
+
" y=time.time()\n",
|
| 640 |
+
" acc.append(test_acc)\n",
|
| 641 |
+
" t.append(y-x)\n",
|
| 642 |
+
" print(f'Epoch: {epoch:03d}, Train Acc: {train_acc:.4f}, Test Acc: {test_acc:.4f}, Time: {(y-x):.4f} seconds')\n",
|
| 643 |
+
"b=time.time()\n",
|
| 644 |
+
"print(f'TOtal time taken:{(b-a):.4f} seconds')"
|
| 645 |
+
]
|
| 646 |
+
},
|
| 647 |
+
{
|
| 648 |
+
"cell_type": "code",
|
| 649 |
+
"execution_count": 27,
|
| 650 |
+
"id": "e160cdc1",
|
| 651 |
+
"metadata": {},
|
| 652 |
+
"outputs": [
|
| 653 |
+
{
|
| 654 |
+
"name": "stdout",
|
| 655 |
+
"output_type": "stream",
|
| 656 |
+
"text": [
|
| 657 |
+
"Epoch: 001, Train Acc: 0.7834, Test Acc: 0.7799, Time: 20.5345 seconds\n",
|
| 658 |
+
"Epoch: 002, Train Acc: 0.8031, Test Acc: 0.7999, Time: 20.3483 seconds\n",
|
| 659 |
+
"Epoch: 003, Train Acc: 0.8063, Test Acc: 0.8047, Time: 20.4774 seconds\n",
|
| 660 |
+
"Epoch: 004, Train Acc: 0.7843, Test Acc: 0.7820, Time: 20.6039 seconds\n",
|
| 661 |
+
"Epoch: 005, Train Acc: 0.8096, Test Acc: 0.8084, Time: 20.7035 seconds\n",
|
| 662 |
+
"Epoch: 006, Train Acc: 0.8012, Test Acc: 0.7979, Time: 20.8724 seconds\n",
|
| 663 |
+
"Epoch: 007, Train Acc: 0.8301, Test Acc: 0.8248, Time: 20.9738 seconds\n",
|
| 664 |
+
"Epoch: 008, Train Acc: 0.7965, Test Acc: 0.7953, Time: 21.0458 seconds\n",
|
| 665 |
+
"Epoch: 009, Train Acc: 0.8086, Test Acc: 0.8059, Time: 21.1957 seconds\n",
|
| 666 |
+
"Epoch: 010, Train Acc: 0.8270, Test Acc: 0.8240, Time: 21.0377 seconds\n",
|
| 667 |
+
"Epoch: 011, Train Acc: 0.8156, Test Acc: 0.8117, Time: 21.0017 seconds\n",
|
| 668 |
+
"Epoch: 012, Train Acc: 0.8316, Test Acc: 0.8289, Time: 21.0952 seconds\n",
|
| 669 |
+
"Epoch: 013, Train Acc: 0.7880, Test Acc: 0.7836, Time: 21.0399 seconds\n",
|
| 670 |
+
"Epoch: 014, Train Acc: 0.8383, Test Acc: 0.8301, Time: 21.0658 seconds\n",
|
| 671 |
+
"Epoch: 015, Train Acc: 0.8307, Test Acc: 0.8258, Time: 21.0292 seconds\n",
|
| 672 |
+
"Epoch: 016, Train Acc: 0.8166, Test Acc: 0.8103, Time: 21.0862 seconds\n",
|
| 673 |
+
"Epoch: 017, Train Acc: 0.8419, Test Acc: 0.8398, Time: 21.0714 seconds\n",
|
| 674 |
+
"Epoch: 018, Train Acc: 0.8011, Test Acc: 0.7987, Time: 21.1226 seconds\n",
|
| 675 |
+
"Epoch: 019, Train Acc: 0.8289, Test Acc: 0.8227, Time: 21.1413 seconds\n",
|
| 676 |
+
"Epoch: 020, Train Acc: 0.8518, Test Acc: 0.8485, Time: 21.1090 seconds\n",
|
| 677 |
+
"Epoch: 021, Train Acc: 0.8334, Test Acc: 0.8294, Time: 21.0528 seconds\n",
|
| 678 |
+
"Epoch: 022, Train Acc: 0.8120, Test Acc: 0.8129, Time: 21.0969 seconds\n",
|
| 679 |
+
"Epoch: 023, Train Acc: 0.7796, Test Acc: 0.7753, Time: 21.0279 seconds\n",
|
| 680 |
+
"Epoch: 024, Train Acc: 0.8334, Test Acc: 0.8309, Time: 21.0795 seconds\n",
|
| 681 |
+
"Epoch: 025, Train Acc: 0.8322, Test Acc: 0.8260, Time: 21.1273 seconds\n",
|
| 682 |
+
"Epoch: 026, Train Acc: 0.8368, Test Acc: 0.8294, Time: 21.1004 seconds\n",
|
| 683 |
+
"Epoch: 027, Train Acc: 0.8486, Test Acc: 0.8403, Time: 21.0755 seconds\n",
|
| 684 |
+
"Epoch: 028, Train Acc: 0.8426, Test Acc: 0.8352, Time: 21.1000 seconds\n",
|
| 685 |
+
"Epoch: 029, Train Acc: 0.8409, Test Acc: 0.8317, Time: 21.0839 seconds\n",
|
| 686 |
+
"Epoch: 030, Train Acc: 0.8315, Test Acc: 0.8276, Time: 21.0667 seconds\n",
|
| 687 |
+
"Epoch: 031, Train Acc: 0.8310, Test Acc: 0.8259, Time: 21.0606 seconds\n",
|
| 688 |
+
"Epoch: 032, Train Acc: 0.8423, Test Acc: 0.8378, Time: 21.0881 seconds\n",
|
| 689 |
+
"Epoch: 033, Train Acc: 0.8564, Test Acc: 0.8525, Time: 21.0627 seconds\n",
|
| 690 |
+
"Epoch: 034, Train Acc: 0.8418, Test Acc: 0.8372, Time: 21.2976 seconds\n",
|
| 691 |
+
"Epoch: 035, Train Acc: 0.8245, Test Acc: 0.8176, Time: 21.2908 seconds\n",
|
| 692 |
+
"Epoch: 036, Train Acc: 0.8377, Test Acc: 0.8315, Time: 21.3887 seconds\n",
|
| 693 |
+
"Epoch: 037, Train Acc: 0.8294, Test Acc: 0.8248, Time: 21.3492 seconds\n",
|
| 694 |
+
"Epoch: 038, Train Acc: 0.8311, Test Acc: 0.8249, Time: 21.3444 seconds\n",
|
| 695 |
+
"Epoch: 039, Train Acc: 0.8646, Test Acc: 0.8570, Time: 21.3087 seconds\n",
|
| 696 |
+
"Epoch: 040, Train Acc: 0.8524, Test Acc: 0.8457, Time: 21.3876 seconds\n",
|
| 697 |
+
"Epoch: 041, Train Acc: 0.8398, Test Acc: 0.8322, Time: 21.3600 seconds\n",
|
| 698 |
+
"Epoch: 042, Train Acc: 0.8227, Test Acc: 0.8204, Time: 21.3573 seconds\n",
|
| 699 |
+
"Epoch: 043, Train Acc: 0.8645, Test Acc: 0.8574, Time: 21.3399 seconds\n",
|
| 700 |
+
"Epoch: 044, Train Acc: 0.8458, Test Acc: 0.8381, Time: 21.3396 seconds\n",
|
| 701 |
+
"Epoch: 045, Train Acc: 0.8712, Test Acc: 0.8624, Time: 21.4001 seconds\n",
|
| 702 |
+
"Epoch: 046, Train Acc: 0.8596, Test Acc: 0.8474, Time: 21.3271 seconds\n",
|
| 703 |
+
"Epoch: 047, Train Acc: 0.8643, Test Acc: 0.8560, Time: 21.4970 seconds\n",
|
| 704 |
+
"Epoch: 048, Train Acc: 0.8583, Test Acc: 0.8511, Time: 21.3742 seconds\n",
|
| 705 |
+
"Epoch: 049, Train Acc: 0.8456, Test Acc: 0.8427, Time: 21.3583 seconds\n",
|
| 706 |
+
"TOtal time taken:3346.5106 seconds\n"
|
| 707 |
+
]
|
| 708 |
+
}
|
| 709 |
+
],
|
| 710 |
+
"source": [
|
| 711 |
+
"for epoch in range(1, 50):\n",
|
| 712 |
+
" x=time.time()\n",
|
| 713 |
+
" train()\n",
|
| 714 |
+
" train_acc = test(train_loader)\n",
|
| 715 |
+
" test_acc = test(test_loader)\n",
|
| 716 |
+
" y=time.time()\n",
|
| 717 |
+
" acc.append(test_acc)\n",
|
| 718 |
+
" t.append(y-x)\n",
|
| 719 |
+
" print(f'Epoch: {epoch:03d}, Train Acc: {train_acc:.4f}, Test Acc: {test_acc:.4f}, Time: {(y-x):.4f} seconds')\n",
|
| 720 |
+
"b=time.time()\n",
|
| 721 |
+
"print(f'TOtal time taken:{(b-a):.4f} seconds')"
|
| 722 |
+
]
|
| 723 |
+
},
|
| 724 |
+
{
|
| 725 |
+
"cell_type": "code",
|
| 726 |
+
"execution_count": 28,
|
| 727 |
+
"id": "b5273211",
|
| 728 |
+
"metadata": {},
|
| 729 |
+
"outputs": [
|
| 730 |
+
{
|
| 731 |
+
"name": "stdout",
|
| 732 |
+
"output_type": "stream",
|
| 733 |
+
"text": [
|
| 734 |
+
"Epoch: 001, Train Acc: 0.8294, Test Acc: 0.8244, Time: 21.6335 seconds\n",
|
| 735 |
+
"Epoch: 002, Train Acc: 0.8411, Test Acc: 0.8347, Time: 21.9815 seconds\n",
|
| 736 |
+
"Epoch: 003, Train Acc: 0.8580, Test Acc: 0.8525, Time: 22.3066 seconds\n",
|
| 737 |
+
"Epoch: 004, Train Acc: 0.8512, Test Acc: 0.8456, Time: 21.8491 seconds\n",
|
| 738 |
+
"Epoch: 005, Train Acc: 0.8688, Test Acc: 0.8576, Time: 21.3336 seconds\n",
|
| 739 |
+
"Epoch: 006, Train Acc: 0.8604, Test Acc: 0.8543, Time: 21.8262 seconds\n",
|
| 740 |
+
"Epoch: 007, Train Acc: 0.8311, Test Acc: 0.8282, Time: 21.4122 seconds\n",
|
| 741 |
+
"Epoch: 008, Train Acc: 0.8623, Test Acc: 0.8533, Time: 21.7136 seconds\n",
|
| 742 |
+
"Epoch: 009, Train Acc: 0.8545, Test Acc: 0.8516, Time: 21.7051 seconds\n",
|
| 743 |
+
"Epoch: 010, Train Acc: 0.8580, Test Acc: 0.8478, Time: 21.6664 seconds\n",
|
| 744 |
+
"Epoch: 011, Train Acc: 0.8400, Test Acc: 0.8350, Time: 22.2604 seconds\n",
|
| 745 |
+
"Epoch: 012, Train Acc: 0.8523, Test Acc: 0.8471, Time: 22.7480 seconds\n",
|
| 746 |
+
"Epoch: 013, Train Acc: 0.8662, Test Acc: 0.8597, Time: 21.5961 seconds\n",
|
| 747 |
+
"Epoch: 014, Train Acc: 0.8509, Test Acc: 0.8452, Time: 21.8072 seconds\n",
|
| 748 |
+
"Epoch: 015, Train Acc: 0.8710, Test Acc: 0.8648, Time: 22.6861 seconds\n",
|
| 749 |
+
"Epoch: 016, Train Acc: 0.8372, Test Acc: 0.8322, Time: 21.4229 seconds\n",
|
| 750 |
+
"Epoch: 017, Train Acc: 0.8435, Test Acc: 0.8354, Time: 21.4031 seconds\n",
|
| 751 |
+
"Epoch: 018, Train Acc: 0.8770, Test Acc: 0.8671, Time: 21.4051 seconds\n",
|
| 752 |
+
"Epoch: 019, Train Acc: 0.8349, Test Acc: 0.8281, Time: 21.3081 seconds\n",
|
| 753 |
+
"Epoch: 020, Train Acc: 0.8362, Test Acc: 0.8324, Time: 21.2875 seconds\n",
|
| 754 |
+
"Epoch: 021, Train Acc: 0.8589, Test Acc: 0.8511, Time: 21.6692 seconds\n",
|
| 755 |
+
"Epoch: 022, Train Acc: 0.8708, Test Acc: 0.8646, Time: 21.2015 seconds\n",
|
| 756 |
+
"Epoch: 023, Train Acc: 0.8649, Test Acc: 0.8575, Time: 21.3023 seconds\n",
|
| 757 |
+
"Epoch: 024, Train Acc: 0.8543, Test Acc: 0.8530, Time: 21.2712 seconds\n",
|
| 758 |
+
"Epoch: 025, Train Acc: 0.8542, Test Acc: 0.8491, Time: 21.0625 seconds\n",
|
| 759 |
+
"Epoch: 026, Train Acc: 0.8427, Test Acc: 0.8381, Time: 21.1436 seconds\n",
|
| 760 |
+
"Epoch: 027, Train Acc: 0.8594, Test Acc: 0.8523, Time: 21.2964 seconds\n",
|
| 761 |
+
"Epoch: 028, Train Acc: 0.8624, Test Acc: 0.8580, Time: 21.2751 seconds\n",
|
| 762 |
+
"Epoch: 029, Train Acc: 0.8614, Test Acc: 0.8568, Time: 21.3692 seconds\n",
|
| 763 |
+
"Epoch: 030, Train Acc: 0.8630, Test Acc: 0.8525, Time: 21.1976 seconds\n",
|
| 764 |
+
"Epoch: 031, Train Acc: 0.8545, Test Acc: 0.8501, Time: 21.0957 seconds\n",
|
| 765 |
+
"Epoch: 032, Train Acc: 0.8444, Test Acc: 0.8355, Time: 21.1121 seconds\n",
|
| 766 |
+
"Epoch: 033, Train Acc: 0.8571, Test Acc: 0.8521, Time: 21.9765 seconds\n",
|
| 767 |
+
"Epoch: 034, Train Acc: 0.8809, Test Acc: 0.8700, Time: 21.5728 seconds\n",
|
| 768 |
+
"Epoch: 035, Train Acc: 0.8607, Test Acc: 0.8502, Time: 21.7194 seconds\n",
|
| 769 |
+
"Epoch: 036, Train Acc: 0.8479, Test Acc: 0.8417, Time: 21.4821 seconds\n",
|
| 770 |
+
"Epoch: 037, Train Acc: 0.8483, Test Acc: 0.8434, Time: 21.6710 seconds\n",
|
| 771 |
+
"Epoch: 038, Train Acc: 0.8621, Test Acc: 0.8519, Time: 21.8791 seconds\n",
|
| 772 |
+
"Epoch: 039, Train Acc: 0.8502, Test Acc: 0.8463, Time: 21.6123 seconds\n",
|
| 773 |
+
"Epoch: 040, Train Acc: 0.8522, Test Acc: 0.8479, Time: 21.8715 seconds\n",
|
| 774 |
+
"Epoch: 041, Train Acc: 0.8525, Test Acc: 0.8450, Time: 21.4852 seconds\n",
|
| 775 |
+
"Epoch: 042, Train Acc: 0.8576, Test Acc: 0.8492, Time: 21.4420 seconds\n",
|
| 776 |
+
"Epoch: 043, Train Acc: 0.8381, Test Acc: 0.8347, Time: 21.7751 seconds\n",
|
| 777 |
+
"Epoch: 044, Train Acc: 0.8674, Test Acc: 0.8605, Time: 21.4741 seconds\n",
|
| 778 |
+
"Epoch: 045, Train Acc: 0.8588, Test Acc: 0.8540, Time: 21.4491 seconds\n",
|
| 779 |
+
"Epoch: 046, Train Acc: 0.8491, Test Acc: 0.8400, Time: 21.4769 seconds\n",
|
| 780 |
+
"Epoch: 047, Train Acc: 0.8439, Test Acc: 0.8376, Time: 21.4490 seconds\n",
|
| 781 |
+
"Epoch: 048, Train Acc: 0.8827, Test Acc: 0.8708, Time: 21.4509 seconds\n",
|
| 782 |
+
"Epoch: 049, Train Acc: 0.8777, Test Acc: 0.8652, Time: 21.4486 seconds\n",
|
| 783 |
+
"TOtal time taken:5775.3160 seconds\n"
|
| 784 |
+
]
|
| 785 |
+
}
|
| 786 |
+
],
|
| 787 |
+
"source": [
|
| 788 |
+
"for epoch in range(1, 50):\n",
|
| 789 |
+
" x=time.time()\n",
|
| 790 |
+
" train()\n",
|
| 791 |
+
" train_acc = test(train_loader)\n",
|
| 792 |
+
" test_acc = test(test_loader)\n",
|
| 793 |
+
" y=time.time()\n",
|
| 794 |
+
" acc.append(test_acc)\n",
|
| 795 |
+
" t.append(y-x)\n",
|
| 796 |
+
" print(f'Epoch: {epoch:03d}, Train Acc: {train_acc:.4f}, Test Acc: {test_acc:.4f}, Time: {(y-x):.4f} seconds')\n",
|
| 797 |
+
"b=time.time()\n",
|
| 798 |
+
"print(f'TOtal time taken:{(b-a):.4f} seconds')"
|
| 799 |
+
]
|
| 800 |
+
},
|
| 801 |
+
{
|
| 802 |
+
"cell_type": "code",
|
| 803 |
+
"execution_count": null,
|
| 804 |
+
"id": "5052b422",
|
| 805 |
+
"metadata": {},
|
| 806 |
+
"outputs": [],
|
| 807 |
+
"source": [
|
| 808 |
+
"for epoch in range(1, 50):\n",
|
| 809 |
+
" x=time.time()\n",
|
| 810 |
+
" train()\n",
|
| 811 |
+
" train_acc = test(train_loader)\n",
|
| 812 |
+
" test_acc = test(test_loader)\n",
|
| 813 |
+
" y=time.time()\n",
|
| 814 |
+
" acc.append(test_acc)\n",
|
| 815 |
+
" t.append(y-x)\n",
|
| 816 |
+
" print(f'Epoch: {epoch:03d}, Train Acc: {train_acc:.4f}, Test Acc: {test_acc:.4f}, Time: {(y-x):.4f} seconds')\n",
|
| 817 |
+
"b=time.time()\n",
|
| 818 |
+
"print(f'TOtal time taken:{(b-a):.4f} seconds')"
|
| 819 |
+
]
|
| 820 |
+
}
|
| 821 |
+
],
|
| 822 |
+
"metadata": {
|
| 823 |
+
"kernelspec": {
|
| 824 |
+
"display_name": "Project",
|
| 825 |
+
"language": "python",
|
| 826 |
+
"name": "python3"
|
| 827 |
+
},
|
| 828 |
+
"language_info": {
|
| 829 |
+
"codemirror_mode": {
|
| 830 |
+
"name": "ipython",
|
| 831 |
+
"version": 3
|
| 832 |
+
},
|
| 833 |
+
"file_extension": ".py",
|
| 834 |
+
"mimetype": "text/x-python",
|
| 835 |
+
"name": "python",
|
| 836 |
+
"nbconvert_exporter": "python",
|
| 837 |
+
"pygments_lexer": "ipython3",
|
| 838 |
+
"version": "3.11.9"
|
| 839 |
+
}
|
| 840 |
+
},
|
| 841 |
+
"nbformat": 4,
|
| 842 |
+
"nbformat_minor": 5
|
| 843 |
+
}
|
requirements.txt
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
ase==3.25.0
|
| 2 |
+
asttokens==3.0.0
|
| 3 |
+
attrs==25.3.0
|
| 4 |
+
catboost==1.2.8
|
| 5 |
+
certifi==2025.1.31
|
| 6 |
+
charset-normalizer==3.4.1
|
| 7 |
+
colorama==0.4.6
|
| 8 |
+
comm==0.2.2
|
| 9 |
+
contourpy==1.3.2
|
| 10 |
+
cycler==0.12.1
|
| 11 |
+
debugpy==1.8.14
|
| 12 |
+
decorator==5.2.1
|
| 13 |
+
executing==2.2.0
|
| 14 |
+
fastjsonschema==2.21.1
|
| 15 |
+
filelock==3.13.1
|
| 16 |
+
fonttools==4.57.0
|
| 17 |
+
fsspec==2024.6.1
|
| 18 |
+
-e git+https://github.com/rampasek/GraphGPS.git@28015707cbab7f8ad72bed0ee872d068ea59c94b#egg=graphgps
|
| 19 |
+
graphviz==0.20.3
|
| 20 |
+
idna==3.10
|
| 21 |
+
ipykernel==6.29.5
|
| 22 |
+
ipython==9.1.0
|
| 23 |
+
ipython_pygments_lexers==1.1.1
|
| 24 |
+
ipywidgets==8.1.6
|
| 25 |
+
jedi==0.19.2
|
| 26 |
+
Jinja2==3.1.4
|
| 27 |
+
joblib==1.4.2
|
| 28 |
+
jsonschema==4.23.0
|
| 29 |
+
jsonschema-specifications==2025.4.1
|
| 30 |
+
jupyter_client==8.6.3
|
| 31 |
+
jupyter_core==5.7.2
|
| 32 |
+
jupyterlab_widgets==3.0.14
|
| 33 |
+
kiwisolver==1.4.8
|
| 34 |
+
littleutils==0.2.4
|
| 35 |
+
llvmlite==0.44.0
|
| 36 |
+
MarkupSafe==2.1.5
|
| 37 |
+
matplotlib==3.10.1
|
| 38 |
+
matplotlib-inline==0.1.7
|
| 39 |
+
mpmath==1.3.0
|
| 40 |
+
narwhals==1.36.0
|
| 41 |
+
nbformat==5.10.4
|
| 42 |
+
nest-asyncio==1.6.0
|
| 43 |
+
networkx==3.3
|
| 44 |
+
numba==0.61.2
|
| 45 |
+
numpy==2.2.5
|
| 46 |
+
ogb==1.3.6
|
| 47 |
+
outdated==0.2.2
|
| 48 |
+
packaging==25.0
|
| 49 |
+
pandas==2.2.3
|
| 50 |
+
parso==0.8.4
|
| 51 |
+
pillow==11.2.1
|
| 52 |
+
platformdirs==4.3.7
|
| 53 |
+
plotly==6.0.1
|
| 54 |
+
prompt_toolkit==3.0.51
|
| 55 |
+
psutil==7.0.0
|
| 56 |
+
pure_eval==0.2.3
|
| 57 |
+
Pygments==2.19.1
|
| 58 |
+
pynndescent==0.5.13
|
| 59 |
+
pynverse==0.1.4.6
|
| 60 |
+
pyparsing==3.2.3
|
| 61 |
+
python-dateutil==2.9.0.post0
|
| 62 |
+
pytz==2025.2
|
| 63 |
+
pywin32==310
|
| 64 |
+
pyzmq==26.4.0
|
| 65 |
+
rascal==0.3.10
|
| 66 |
+
referencing==0.36.2
|
| 67 |
+
requests==2.32.3
|
| 68 |
+
rpds-py==0.24.0
|
| 69 |
+
scikit-learn==1.6.1
|
| 70 |
+
scipy==1.15.2
|
| 71 |
+
seaborn==0.13.2
|
| 72 |
+
six==1.17.0
|
| 73 |
+
stack-data==0.6.3
|
| 74 |
+
sympy==1.13.3
|
| 75 |
+
threadpoolctl==3.6.0
|
| 76 |
+
torch==2.4.1+cu121
|
| 77 |
+
torch_cluster==1.6.3+pt24cu121
|
| 78 |
+
torch_geometric==2.3.1
|
| 79 |
+
torch_scatter==2.1.2+pt24cu121
|
| 80 |
+
torch_sparse==0.6.18+pt24cu121
|
| 81 |
+
torch_spline_conv==1.2.2+pt24cu121
|
| 82 |
+
tornado==6.4.2
|
| 83 |
+
tqdm==4.67.1
|
| 84 |
+
traitlets==5.14.3
|
| 85 |
+
typing_extensions==4.13.2
|
| 86 |
+
tzdata==2025.2
|
| 87 |
+
umap==0.1.1
|
| 88 |
+
umap-learn==0.5.7
|
| 89 |
+
urllib3==2.4.0
|
| 90 |
+
wcwidth==0.2.13
|
| 91 |
+
widgetsnbextension==4.0.14
|
san_best.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0dd7b736400736375f93e36fe2e5150e2c21258a709d998b028dd6900fd112fc
|
| 3 |
+
size 757556
|
test_classicml.ipynb
ADDED
|
File without changes
|
transformer_classifier_relu.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c76651ae52ee11de9167c281960117bd01c0d5b6a326dc5fef74caeff14ade83
|
| 3 |
+
size 1540122
|
tsne_data.csv
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|