Recently Browsing 0 members
No registered users viewing this page.
The function below worked for a long time. Now it doesn't download anything and I don't know where the error is.
#include <InetConstants.au3> Local $Link = 'https://allegro.pl/kategoria/nasiona-warzywa-99776?bmatch=e2101-d3681-c3682-hou-1-4-0319' Local $dStrona = InetRead($Link) Local $sStrona = BinaryToString($dStrona, $SB_UTF8) ;GUICtrlSetData($myedit, $sStrona) ConsoleWrite($sStrona) The function was good because it reading data in the background.
Thank you for any hints or possibly other solutions / functions that are reading data in the background.
edit: this is probably in the wrong place, can a moderator move it to wherever it belongs?
is there any way to completely disable TCPTimeout and make TCPRecv() wait indefinitely? maybe setting it to 0 or -1 or something?
the docs doesn't seem to mention any way to disable it
- the underlying C code would set SO_RCVTIMEO to 0 , aka
setsockopt(s, SOL_SOCKET, SO_RCVTIMEO, &timeout, sizeof(timeout));
Here test example:
The code is not prefect, but it's demonstrates the problem.
So, when I download more than 2 files simultaneous from one server (in my example I download 3 simultaneously) by InetGet() with background option, I will get only 2 active downloads. It's seems like server limitation. The third will remain with 0 read bytes until one of the previous downloads would finished. And if I want to close this third 0 bytes read download with InetClose() I will get hang of Window GUI for several seconds! After that, the third download would be closed and everything will continue to work normally.
If I InetClose() one of the active downloads with more than 0 bites read - all works fine!
You can try by your self given example, or watch this .gif's:
1. Window GUI hangs when I trying to Cancel (InetClose()) download with 0 bytes read download:
2. All works as it should, when I Cancel more than 0 bytes read download:
Any ideas why is it happens?
any suggestion to track down, why a script is freezing after a longer run time?
This is a general question, I've checked the common mistakes, I've made so far (inside a loop doing fileopen() without fileclose(), runwait() for progs, that don't terminate, infinite recursive loops etc.)
This script stops just after several days of running...
Is there a direct way to "count" the amount of MEM and handles, a autoit script is using itself, or by child processes?
Hi everyone, hope you are doing fine
Well, I'm currently writing a small script that goes to a certain web page, finds the first link of a specified section and download the file associated to this link.
Depending on the computer that the tool is launched, the script gets the computer model and search in the (provided here) ini file which link to follow.
At first, Dell was kind enough to provide only one link but now, they provide two of them. The first one is now a .txt file ( ) whereas my script has been designed to download only the fist and latest link released for the BIOS Update.
Here's the current code which is working with only the first and latest link of the BIOS category:
So the question is:
In the case of double links like shown in the picture above, how it is possible to tell the script to download only the link containing an the .exe file?
Of course, I could have changed the array result to  instead of  [which is working] but it seems that Dell does that randomly and that I deal with a lot of computer models.
Thanks for the help you can provide,