Response.IsClientConnected Property As A Performance Optimization For .net Websites?
Aug 17, 2010
I was wondering if anyone had experience they could share using the Response.IsClientConnected property as a performance optimization for asp.net websites.
The reason I ask is that I am a bit skeptical on how effective it would be in real life scenarios. I understand the concept of checking the value before performing a large task but I just can't see how useful this would be as clients could disconnect at any point time.
I have database with 40 big tables.In each stored procedure query is containing minimum 6 joins.The query is taking too much time to execute.I didn't implemented views and indexes.
When my Application face a long-time process, i.e fetch a query (SELECT a, b, c FROM d)This query needs 10 seconds to be completed in the MSSQL Management Studio, but when the ASP.NET application try to fetch it, it refuse to return any response to any other requests made on that Server.
I am hosting my Application on VPS Server with good specifications, and I am giving this example the (SELECT a, b, c FROM d) just to tell you the issue, it can be any process, maybe processing a movie, or even fetching some data through external API that is experiencing some slow-down,or whatever.
I want to implement a response time monitor in a ASP.Net application. This monitor has to calculate the average response time during 2 secs. Whats the performance counter i can use. I tried ASP.NET ApplicationsRequests/Sec but did not work. Whats the best way to do this. I am using the default server formed by VS2005 for development, does this affect the performance counter programming?
Where should I implement the logic? Global.ASAX? in every 2 sec i want to check this value and make some decisions.
In addition, what are the parameters I can alter in a ASP.Net application at runtime to improve perfromance. How can I change them?
I have a web performance test which contains a request whose response is greater than 5MB, and the Extract Hidden Fields rule fails to find (necessary and required!) hidden fields in the response. Response header contains
HTTP/1.1 200 OK Transfer-Encoding : chunked Vary : Accept-Encoding, User-Agent Cache-Control : private Content-Type : text/plain; charset=utf-8 Date : Sat, 19 Feb 2011 15:24:38 GMT Server : Microsoft-IIS/6.0 X-AspNet-Version : 2.0.50727 X-Powered-By : ASP.NET
Other than that and the response size, there is nothing remarkable about this scenario. In fact, this same test succeeds when a smaller data set is used. I suspect the Web Performance Test framework is having issues parsing the "chunked" encoding or sheer volume of data. Ahem, how can I obtain these required hidden fields from my response? ie resolutions, work arounds, converting auto-extraction to manual, etc.
I have question about HttpContext class (http://msdn.microsoft.com/en-us/library/system.web.httpcontext.aspx). This class has Responseproperty (http://msdn.microsoft.com/en-us/library/system.web.httpcontext.response.aspx). But I don't understand, why Response property hasHttpResponse return value, not HttpContext (this is HttpContext, not HttpResponse property).So, why there is:
the account we use to copy/publish websites to our web server is not allowing us access to push websites. I've added this account to the permissions in the links listed in the below link, but i don't have success when trying to push remotely. When the account is in the local admins account, everything works flawlessly, but we are not allowed to have it in there. http://social.msdn.microsoft.com/Forums/en-US/vssetup/thread/31be047e-4716-4974-b8a1-be0111b50199 I've googled and searched a lot for this particular error, but am not finding an answer that helps. We get this error, 'Unable to create the Web 'http://edea01/test/planning'. You are not authorized to perform the current operation', and the above link is the scenario that matches mine the most. I don't do develompment work, but am asked to figure out this connection problem.
I have developed a website in asp.net framework 2 . This website is being hosted in two different servers without any change in code. My issue is about the performance of these 2 sites. One website is taking much time for inserting datas to the DB (SQL server 2005). 2 websites are having different DB server.
I think the issue is for the DB server. How can we rectify the DB performance while insertion and Is there any other cause for this permance issue?
I am not sure if this is the right forum. I can not find a forum for LINQ.
I am working on an application using LINQ. Application performance is not up to par and my tests show that it is LINQ queries that are slow. I was wondering if anybody can recommend where I can find an article about optimizing LINQ performance maybe by compilation or other methods.
I am creating a service oriented application where trying to have everything using services....however there is something I am not sure of , I am having a page that calls the database at the page load...so what would be better and faster?? to call database in pageload , or to call wcf service from javascript during javascript load ??btw , I am using a repeater in the page , but I have created somekind of an engine to create the suitable html so...I'll be creating the repeaters html using the wcf and resend it back to the page If I am using a wcf service at the start.
i have two models that represent two related tables:
class Article { Guid ID; Guid AuthorID; string Title;[code]...
i need to show this view model in a view:
class ArticleWithAuthor { string ArticleTitle;[code]...
in a controller for Article i need to create a view model for the given article id. i read somewhere that a good way to create a view model is to write a mapper that translates from model to view model.
my question is:since the mapper needs a model to work with, it seems that i need two database requests: one to retrieve my Article model and the other one to get my Person data to merge them in the model view. this seems inefficient. is there a way to combine these requests and still follow this "Entity Translation Pattern"? i know of course that i can create my view model "manually" by writing a custom LINQ expression but i am trying to follow the MVC guidelines.
I am currently working on an c# application that minimizes and combines javascript/css asynchronously in the background to load on to the page which completed. Once the combined minimized file(s) are created, they are saved on to disk and all subsequent requests to the page will load this file.
The reason for this is to assist with performance on the page. I have a concern though, what if the combined file is large, eg 200 kb. Would it be better to combine in to 2 files if this was the case and have 2 separate http requests? The file will be gzipped and cached.
I have an application using data from multiple stored procedures in a single database.The dataacess is disconnected one. Entire data is required during loadtime i.e. all calls made at load time. I want to know what is the most optimized way to do it. As the connection is required with only one database,is there a way to create it once and use it for the next times. I am aware of connection pooling but its anyways done by default. kindly suggest some way to optimize the load time. Also the result set from all the procedures cannot be clubbed so multiple stored procedure invoking is compulsory required.
i have awebsite and i made all my steps to make it in the first links on the search of google (SEO) and now i want to go to google updates to make the steps that apply the website on google after one day almost. can any one tell me what i should begin with or do?
I am working on a large project where I have to present efficient way for a user to enter data into a form.
Three of the fields of that form require a value from a subset of a common data source (SQL Table). I used JQuery and JQuery UI to build an autocomplete, which posts to a generic HttpHandler.
Internally the handler uses Linq-to-sql to grab the data required from that specific table. The table has about 10 different columns, and the linq expression uses the SqlMethods.Like() to match the single search term on each of those 10 fields.
The problem is that that table contains some 20K rows. The autocomplete works flawlessly, accept the sheer volume of data introduces deleays, in the vicinity of 6 seconds or so (when debugging on my local machine) before it shows up.
The JqueryUI autocomplete has 0 delay, queries on the 3 key, and the result of the post is made in a Facebook style multi-row selectable options. (I almost had to rewrite the autocomplete plugin...).
So the problem is data vs. speed. Any thoughts on how to speed this up? The only two thoughts I had were to cache the data (How/Where?); or use straight up sql data reader for data access?
I have more products in a single aspx page. So I want to upload this page in my remote server where my website is running. How can I test whether search engine can resolve my product on that page?
I am using VS 2010 login form and aspnet table for user creation and role assignment. Every thing is working excellent.
My concern is that more 200 users will logon on this site simultaneously what will be the effect on performance? how to optimize site speed when more than 200 users will be log on.
I am working on a web application that is consuming a Dataset returned by a web service. As the application is running I store that Dataset as a session variable to be used over and over again as the user navigates to the different pages that will edit the tables within the dataset. The idea was the user will only have to wait for the data once when the application loads then the application would use the session variable until the user saves the changes they made, when that happens it would pass the edited tables to the service to update the database. Is there problems with this design and the storage of the Dataset and Datatables as a session variable?
I catch the error of "Property Evaluation Failed" for the sentence "Response.End" in the following code.
Sub exportToExcel(ByVal _DT As DataTable) Dim dt As DataTable = _DT Dim attachment As String = "attachment; filename=userlist.xls" Response.ClearContent() [code]....
I am using this code to download and its working well for me.But i cant understand the code.Can someone explain me this code to me please?
Response.AddHeader is used to add a new HTML header,but what is an HTML header all about?and the parameters i am passing within it as the name and value;what are they?
Even though I have buffer set to false,no text is displayed.I have tried adding a response.flush,with and without changing the buffer value.What exactly is wrong?I've also tried it with and without the label(i.e. with or without just Response.Write)