@@ -89,28 +89,28 @@ async def handle_connection(self, reader: asyncio.StreamReader,writer: asyncio.S
8989 return
9090 node_id = data [4 :]
9191 with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
92- log .write (f"connection is from { addr } { node_id } \n " )
92+ log .write (f"connection is from { addr } { self . get_peer ( node_id ). pub_key } \n " )
9393 if self .locks .get (node_id ) == None :
9494 self .locks [node_id ] = asyncio .Lock ()
9595 # print("connection from",node_id)
9696 if self .connections .get (node_id ) != None :
9797
9898 with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
99- log .write (f"duplicate connection from { addr } { node_id } \n " )
99+ log .write (f"duplicate connection from { addr } { self . get_peer ( node_id ). pub_key } \n " )
100100
101101 if self .connections .get (node_id ).opened_by_me * ternary_comparison (self .peer .id_node , node_id ) == - 1 :
102102 with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
103- log .write (f"closing previous with { addr } { node_id } \n " )
103+ log .write (f"closing previous with { addr } { self . get_peer ( node_id ). pub_key } \n " )
104104 self .remove_from_dict (node_id )
105105 else :
106106 with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
107- log .write (f"keeping old one with { addr } { node_id } \n " )
107+ log .write (f"keeping old one with { addr } { self . get_peer ( node_id ). pub_key } \n " )
108108 writer .close ()
109109 return
110110 writer .write (int (0 ).to_bytes (32 ,byteorder = "big" ))
111111 await writer .drain ()
112112 with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
113- log .write (f"listening from { addr } { node_id } \n " )
113+ log .write (f"listening from { addr } { self . get_peer ( node_id ). pub_key } \n " )
114114 self .connections [node_id ] = DictItem (reader ,writer ,None , - 1 )
115115 self .connections [node_id ].unique_id = urandom (4 )
116116 self .connections [node_id ].confirmed = True
@@ -271,9 +271,9 @@ async def listen_for_data(self, reader: asyncio.StreamReader, node_id = None, ad
271271 buffer = bytearray ()
272272 i = int .from_bytes (data ,byteorder = "big" )
273273 if i != 0 :
274- # with open(f"log{self.peer.pub_key}.txt", "a") as log:
275- # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
276- # log.write(f" will from {self.get_peer(node_id).pub_key} {i} {len(data)}\n")
274+ with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
275+ # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
276+ log .write (f" will from { self .get_peer (node_id ).pub_key } { i } { len (data )} \n " )
277277
278278 while i > 0 :
279279 data = await reader .read (min (i , 9048 ))
@@ -307,29 +307,32 @@ async def send_stream(self, node_id, data, lvl = 0):
307307 return False
308308 try :
309309 async with self .locks [node_id ]:
310- # with open(f"log{self.peer.pub_key}.txt", "a") as log:
311- # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
312- # log.write(f" sending to {self.get_peer(node_id).pub_key} {len(data)}\n")
310+ with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
311+ # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
312+ log .write (f" sending to { self .get_peer (node_id ).pub_key } { len (data )} \n " )
313313
314314 self .connections [node_id ].writer .write (len (data ).to_bytes (32 ,byteorder = "big" ))
315315 await self .connections [node_id ].writer .drain ()
316316 self .connections [node_id ].writer .write (data )
317317 await self .connections [node_id ].writer .drain ()
318318 except ConnectionResetError :
319- # with open(f"log{self.peer.pub_key}.txt", "a") as log:
320- # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
321- # log.write(f" cannot send to {self.get_peer(node_id).pub_key} {len(data)}\n")
319+ with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
320+ # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
321+ log .write (f" cannot send to { self .get_peer (node_id ).pub_key } { len (data )} \n " )
322322 await asyncio .sleep (3 )
323323 p : Peer = self .get_peer (node_id )
324324 if p == None :
325325 return False
326326 ret = await self .open_connection (p .addr [0 ],p .tcp , p .id_node , port_listen = 0 )
327327 if ret == False :
328328 return False
329+ with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
330+ # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
331+ log .write (f"Resetting sending to { node_id } \n " )
329332 return await self .send_stream (node_id ,data , lvl = lvl + 1 )
330- # with open(f"log{self.peer.pub_key}.txt", "a") as log:
331- # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
332- # log.write(f" finished sending to {self.get_peer(node_id).pub_key} {len(data)}\n")
333+ with open (f"log{ self .peer .pub_key } .txt" , "a" ) as log :
334+ # log.write(datetime.now().strftime("%d/%m/%Y, %H:%M:%S"))
335+ log .write (f" finished sending to { self .get_peer (node_id ).pub_key } { len (data )} \n " )
333336 # print("done srream")
334337 return True
335338 def set_stream_close_callback (self , callback ):
0 commit comments