I'm using a third party url to determine the location of site visitors based on IP address.I use WebClient.DownloadData() to get the information returned by the url which is a one word answer e.g. as per http://api.hostip.info/country.phpOnce I have detected what country the user is in I want to either A) keep them on the current site and store the country in the session or B) redirect them to another countries site.What I'm worried about is WebClient.DownloadData() hanging so the user is waiting for ages and can't use the site. I'd thought about using DownloadDataAsync but this dosn't seem to be a solution for me as I don't want them to do anything till I find out the country (also seettp://stackoverflow.com/questions/271767/webclient-downloaddataasync-is-freezing-my-ui)Is there a way I can set a timelimit on my call to WebClient.DOwnloadData e.g. if it hasn't returned in 2 seconds,
I'm mix up between webclient.OpenReadAsync and webclient.DownloadStringAsync? Can anyone explain clearly for me ? What are the difference between them? In addition, may i know whether webclient.OpenReadAsync got download the file or just open and read the file only without download to other places?
We have the timeout value set to 120 in our <form> tag within the web.config. We do not have a session timeout set.. and we have various connection strings.
We are having a problem where a session variable will disappear (become NULL) .. but, the form evidently remains 'open'.. or no re-login is required..... so, my question(s):
1. what is the relationship between form timeout and session timeout
I am having a problem with my ASP.Net Web application. The application is developed using vb.net and is linked to a SQL Server database. Let me explain how the application works and the problem I am experiencing. The system is an online web app which allows registered users to create a CV online. One of the pages within the app gives users the chance to add a cover note to their CV. The page that allows them to do this consists of only a textarea control and a button control. The textarea allows users to input up to 4,000 characters.
Once the user clicks the 'Save' button to save their cover note info, the following code then executes.This code checks to see if the CV already has cover note info, if it does, then the application runs an update statement, otherwise, it runs an insert statement.The table within the database which records the cover note information is called tbl_covernote and has three columns, covernote_id (int and autoincrement), cv_id(int), covernote_text (nvarchar(max)).
The error which occurs sometimes is as follows:
Dim dr As SqlDataReader Dim param(0) As SqlParameter Finally [code]...
I have an ASP.Net application that makes an AJAX request to retrieve at report. The report can run for a long time so I set the asyncpostbacktimeout in <asp:ScriptManager /> to 600. However, when I try to run the report, if it runs for longer than 90 seconds it fails to come back. I can see in the IIS logs that the POST request succeeded with a 200 status and I can see the time taken is much less than 600.
The web page dutifully waits for the entire 600 seconds before returning with a timeout error:
Error:
Sys.WebForms.PageRequestManagerTimeoutException:
The server request timed out.
Is there any setting I should be checking in IIS? Connection timeout is 900 seconds.
I have a page of each every click has ajax call to my server (hence, the ASP extends the session)
I have ASP.NET session set to Xmin. I want when X+1 min expires, I have expiration page. what I did was to set the JS timer to validate every x+1min to see if the session expired (the problem is that the JS and the ASP session timeouts are not synced)
I have an issue to where my users are logged into my system but thier session is null so when they try to do something in the system after 20 minutes, i get null reference expceptions because my session is gon What is the best practice for handleing the session, should i kick the users out before thier session ends and when they log back in thier session will return or atleast a new one will be created right? How should I handle this?
I'm working with ASP.Net web services and am having a problem with a long-running process that takes about 5 minutes to complete, and it's timing out. To fix this, I was able to set the executionTimeout on the server's web.config to 10 minutes, and then set the .Timeout property on the Web Service object to approximately 9 minutes. Now, I'm worried that this may possibly cause some other web service calls to sit there for 10 minutes before they time out rather than the previous 90-100 seconds. I know the default on the client side is 100 seconds, but wasn't sure if updating the server's timeout setting would affect this.
Bottom line is - Is it safe to update the server's timeout setting to a long amount like 10 minutes, and rely on the default timeout on the client, or could this end up causing some problems?
I run a loop that downloads a lot of smal files from another website
try Loopstarts result = runData() loop ends catch ex1 as exception End try private function runData() as string Dim wc as new WebClient Dim str as string try str = wc.downloadstring(filePath) catch ex2 as exception return "Nothing" end try wc.dispose() finally return str end function
THE Problem is if the file is not found in the RunData function I get 'Thread being aborted' error which stops the not only the function RunData() but also the loop is running in, so I get an error in ex2 as well as ex1. But I want the error to be captured in ex1 and stop, and not effect the loop.
I have read that you need to use Response.End to close the thread or not to capture the 'threadbeingaborted' expception.
BUT how do you do the response.end with a webclient??? (IN vb.net , asp.net 2.0 version
I've got a site that users can use to check in and out documents. To check out a document, the user slects a file from a list, and clicks 'check out' . The document is taken from a 3rd party source control tool and placed in a dedicated directory on the server. Then, the file needs to be copied from this directory to a client directory the user has specified.
Copying to the Server is no problem.
I'm trying to use the WebClient objects DownloadFile method to write to the clients directory, but I keep getting "Access to the path '<path>' is denied.". I've set the UseDefaultCredentials flag to true, but no joy.
I'm obviously missing something, but I'm not sure what.
Anyone familiar with the webclient.downloadfile method method? I'm looking for a way to help two users download files from an internal secure website. Right clicking File Save As, and then browsing to a folder is "too many steps". So, I'm curious what this method can do to reduce clicks and/or automate the download process.
I'm using WebClient.DownloadString("http://www.website.com/Default.aspx?fltdte=01050402);Part of data that is returned I want to put again in above url for query again and again if data returned satisfy conditions..I want to do multiple webClient.DownloadString.How to do that?
Error 1 'System.Net.WebClient' does not contain a definition for 'DownloadData' and no extension method 'DownloadData' accepting a first argument of type 'System.Net.WebClient' could be found (are you missing a using directive or an assembly reference?) C:UsersMichaelDocumentsVisual Studio 2008Projectssearch2search2MainPage.xaml.cs'm doing this in a c# file for a XAML/Silverlight project, but can't imagine why that would make a difference. I can not find any reference to this issue on the web, and I had something similar to this working last month, but in a regular ASP.NET page, not in a Silverlight app.
I have an objective to transfer some files which can be pdf, jpg and xml from one location to another. I am creating a handler based API for that. What could be the best approach for doing it? Should i return a byte array so that the client can pick the byte array and convert it back to file? Also the max file size I need to handle is 18 MB, so i have to make sure that the sizes are properly handled. Should I do an asynchronous transfer for that?
Not sure if any of you are familiar with hMailServer, however, I am attempting to automate account creation / editing / deletion by using the WebClient class to interact with hMailServer's PHPWebAdmin.
The problem I am facing is that I can't seem to retrieve the index page (http://localhost/hmailserver/index.php) after posting the login credentials.
Here's my code so far....
[Code]....
And here's the HTML for the login form...
[Code]....
what I'm missing / potential problems? Should the 'url' be the page with the form of which I wish to post data to, or the page I want returned? I've tried setting it to "....hmailserver/index.php" or to the status page ...hmailserver/index.php?page=status" but the response returned is always the HTML of the login page.
I am using the code below to download multiple attachments from a TFS server:
[code]....
However, there are a couple of problems with this approach:
The Thread.Sleep() is locking up my Form. I still need to make my own Thread or use BackgroundWorker. (I would like to avoid this as much as possible)The DownloadFileCompleted event is being triggered after ALL files has been downloaded. I don't know if this is a side-effect of using
System.Threading.Thread.Sleep(1000);
Is there a better approach to download files one at a time using WebClient.DownloadFileAsync?
My question is similar to http://stackoverflow.com/questions/194579/how-to-detect-when-a-user-has-successfully-finished-downloading-a-file-in-php but I have to do the same using IIS, ASP.NET and C#.
None of the methods in the HttpResponse class provide feedback if the data was sent or not, TransmitFile just does its job (or not) and does not provide any means of knowing the result.I was thinking of using the .Filter property but then again, the filter is based on the HttpResponseStream which does also not provide any feedback.