I am working on an ASP.NET 3.5, C#, ASP.NET AJAX, JQUERY web application which currently is having some performance issues. I have a screen with the forms view control and when I edit the form and save it for the first time the POST takes 4.89 seconds and when I perform the same operation again for the second time the POST only takes 1.09 seconds. What could be the reason for this strange behavior?
Note: I am using firebug and I am disregarding the page resources (like js, images) load time.
EDIT: I am using the Web Deployment Project to precompile the application.
I have developed a website in asp.net framework 2 . This website is being hosted in two different servers without any change in code. My issue is about the performance of these 2 sites. One website is taking much time for inserting datas to the DB (SQL server 2005). 2 websites are having different DB server.
I think the issue is for the DB server. How can we rectify the DB performance while insertion and Is there any other cause for this permance issue?
I am not sure if this is the right forum. I can not find a forum for LINQ.
I am working on an application using LINQ. Application performance is not up to par and my tests show that it is LINQ queries that are slow. I was wondering if anybody can recommend where I can find an article about optimizing LINQ performance maybe by compilation or other methods.
I am creating a service oriented application where trying to have everything using services....however there is something I am not sure of , I am having a page that calls the database at the page load...so what would be better and faster?? to call database in pageload , or to call wcf service from javascript during javascript load ??btw , I am using a repeater in the page , but I have created somekind of an engine to create the suitable html so...I'll be creating the repeaters html using the wcf and resend it back to the page If I am using a wcf service at the start.
I'm now getting into EF and from what I'm seeing so far, I wouldn't have to worry about writing stored procedures any more. Looks like EF takes care of all of that -- including INSERT's that store data in multiple tables. One of the things they taught us when learning stored procedures was that they're compiled which has performance benefits. How does EF 4 fare against using stored procedures?
I'm trying to make a decision about how to display my data. What I have now is a list of products displayed in a repeater. But for code-maintenance I've put my product items in a seperate usercontrol and loading them in a loop with db results using LoadControl.The product control itself is very simple, just a few public properties like title, url, rating but I'm not sure if this will affect my performance. I did some reading here and on forums and some people say it's not the best practice especially if you have more then 20 or 30 of these controls. So, is it really a performance hit using this method or does it stay ok with around 10.000 hits a day.
I have developed a gridview which has displays 8,000 records. However, the user is compaining that it is too slow. A major function of this gridview is to filter on two of the columns. It seems to me that if the records were stored in memory it would be much faster. This is caching, I think. Is that the case? Since filtering is being done I don't think custom paging will help much. When they submit a filter, with custom paging wouldn't the entire table be read again? So I was thinking of trying caching which is available on SQL Server. Am I on the right track? Can you cache with an Access database? Can you think of any other ways to improve performance of a gridview?
I did not post this in database forum because the data will run on an asp.net site.I have 3 solution in my mind and i would like some opinions.What i want to do is create a parent-child relationship on 2 controls.A listbox and a textbox.I will populate the listbox with data(this is taken care of so ignore) and what i want to do is when i click on an item in the listbox then the textbox to show it's joined related item(one item only).This is not a problem in windows forms app but in asp, since i'm new i don't know what would be better for a faster retrieval on the page.So the 3 options i have in mind are:
1)DAL.The standard create a query and let one @parameter wait for the id.
2)ADO.NET with possibly an sqldatasource.In standard forms i would have chosen simple ado.net but in asp?So either constantly open,close the db and retrieve the item or use an sqldatasource and chance the parameter.
3)Asynchronous handler page.Bind the textbox to an asynchronous page that contains the connection and expect the @id parameter.
I admit that i'm not fond of DAL but if it will boost speed then i will use it.But i have a though that says that simple ADO will be faster.
I have a strange situation on a production server. Connection for asp.net get queued but the CPU is only at 40%. Also the database runs fine at 30% CPU.
Some more history as requested in the comments:
In the peak hours the sites gets around 20,000 visitors an hour. The site is an asp.net webforms application with a lot of AJAX/POSTs The site uses a lot of User generated content We measure the performance of the site with a testpage which does hit the database and the webservices used by the site. This page get served within a second on normal load. Whe define the application as slow when the request takes more than 4 seconds. From the measurements we can see that the connectiontime is fast, but the processing time is large. We can't pinpoint the slowresponse the a single request, the site runs fine during normal hours but gets slow during peak hours We had a problem that the site was CPU bound (aka running at 100%), we fixed that We also had problems with exceptions maken the appdomain restart, we fixed that do During peak hours I take a look at the asp.net performance counters. We can see behaviour that we have 600 current connections with 500 queued connections. At peak times the CPU is around 40% (which makes me the think that it is not CPU bound) Physical memory is around 60% used At peak times the DatabaseServer CPU is around 30% (which makes me think it is not Database bound)
My conclusion is that something else is stopping the server from handling the requests faster. Possible suspects:
Deadlocks (!syncblk only gives one lock) Disk I/O (checked via sysinternals procesexplorer: 3.5 mB/s) Garbage collection (10~15% during peaks) Network I/O (connect time still low)
To find out what the proces is doing I created to minidumps.
I managed to create two MemoryDumps 20 seconds apart. This is the output of the first:
!threadpool CPU utilization 6% Worker Thread: Total: 95 Running: 72 Idle: 23 MaxLimit: 200 MinLimit: 100 Work Request in Queue: 1 Number of Timers: 64
and the output of the second:
!threadpool CPU utilization 9% Worker Thread: Total: 111 Running: 111 Idle: 0 MaxLimit: 200 MinLimit: 100 Work Request in Queue: 1589
I'm trying to add some performance counters to my asp.net website. Now, I know how to increment/decrement some custom counter I make .. but my problem is that if I get my ASP.NET website to create these counters, if they do not exist (eg. i do this check in the global.asax App start method) then add/create them.
But, it doesn't work - access to the registry is denied/forbidden.
I'm assuming this is because the asp.net process is so stripped down (for security) that u can't touch that type of thing. Therefore, i'm wondering if the only other solution is to make a quick console or winform app which does one thing -> add's the perf counters. running this as my normal logged in user would me i have admin rights, so it will work. or is there something else i can do?
Does it have any major effect on performance/ memory if my web.config is really huge (say, 1000+ entries in <appSettings>)? Is it a good idea to maintain a different custom xml config file for all business specific settings for my app?
I am using ASP.NET AJAX Autocomplete to load the data. I am fetching data from a Dataset, which is store in Session. The table length is approx 1000 rows.I use DataView to and Table.Select ("Data Like '"+ inputvalue +"'") to filter data on each keystroke.I am wondering, what's the best and fastest way to store and fetch data for Autocomplete?
Does Modelbinding increase performance that much? I have a large table which I have split into 4 details/edit screens, and want to load only information that is needed on the particular screen.
My scenario is : WebApp -> WCF Service -> EDMX -> Oracle DB
When I want to bind grid I fetch records from Oracle DB using EDMX i.e LINQ Query. But, this degrades performance as multiple layers take place between WebApp & Oracle DB. Can I use caching mechanism to improve the performance? But as far as I know cache is shared across the whole application. So, if I update cache other user might receive wrong information. Can we use caching per user? Or is there any other way to improve performance of the application?
I know red-gate's ANTS Performance Profiler but it works with source code. I have an ASP.NET web application that is deployed on a remote server. I'm going to test its performance while it is working on the server.
My main goal is to detect what method of what class is slow and then is a bottleneck. Is there a way or a tool?
I support an inherited asp.net web application that retrieves files from a shared network drive. One particular folder [we'll call it FOLDERA] I've chosen to troubleshoot against contains about 300mb of files and multiple subfolders. FOLDERA is considered large for this application. Starting recently, the users have been having slow performance when retrieving files from FOLDERA on a production network share. I narrowed the code down to the GetDirectories method within the asp.net application.
In production, the FOLDERA read takes about 8-10 seconds. Prior to the recent performance degradation it was about 1 second. In the test environment it takes 1-2 seconds with the same amount of data.
I have a pretty big web site (asp.net web) and I have used JavaScript intensively (jquery,custom javascipt, etc) and also The theme and CSS. Right now I have a huge amount of js and css files in my system and I am thinking about minifying and smashing the js and css to improve the performance. So, please advice me the suitable tools and technologies to be used and please suggest me the best practices to be applied in these scenario.
I am working on an ASP.NET project that has significant backend processing existing in PLSQL. I was wondering if I re-write this code properly in C# can I get similar performance results or is the PL/SQL code going to beat my C# code every time?
Web forms .NET application, is starting to meet 64 bits Windows servers and we have a 32bit DLL from one of our component vendors (its a hardware key, actually). I checked the instructions here "compiling a .net application with either a 32-bit or 64-bit dll". It is also what our vendor sugested. question: is there a performance impact on compiling a web application to run on x86 mode ?Please don't flame me for using a hardware key
I find my self in a situation where I have to choose between either creating a new sp in database and create the middle layer code. so loose some precious development time. also the procedure is likely to contain some joins.Or use two already existing sp(s), the problem of this approach is that I am doing two round trips to database. which can be poor performance especially if I have database in another server.
I am a web designer and usually design corporate web sites which often does not require update. So I want to cache the output for one day. How can I do this?
I have an self-build asp.net page, (link), and on my own localhost-server it runs very fast.But when I have uploaded it to my rent server (Win08Server, 3GB Ram, 3GhZ...) it goes very very slow (3-5 second a site-switch).
hot to find out what programm is disturbing or where I have a leak?
(the most of the content is static, only a 2 database querys at the beginning!!!)
Until now, I have not create any massive applications using ASP.Net. However, I am looking to create an application that has the potential to be very performance intensive. So I am looking for some tools or best practices when it comes to performance. I would like to be able to know how to: See my current performance (good or bad) View items that need fixing And being able to compare two performance variable items would be great as well.