I've noticed as my website gets bigger and bigger, the time my laptop takes to display my page is much longer then say a new projects with minimal references. I think there are two variables at play that affect ASP.NET warm-up time:
The quantity of external references The time it takes for a worker process to new() up each instance per worker process Additional time for the WCF objects as the ServiceHost may be in an external DLL First, are those the correct variables to take into account when considering ASP.NET startup time? Next, it appears that web.config may dispatch other objects for use with certain filetypes (*.svc, *.aspx, Windows Identity Foundation (WIF), etc. ). This too may cause delays in ASP.NET.
Last, my project is created as a "web project" not a "web site". Not sure if this has an impact. Is my theory full of holes, or is there something I can do to make development on a old laptop any better?
I've got a div on Page A.When an event happens (whatever I define, a click, whatever), I do a div.Dialog, set its properties and open itThe data inside is returned from an async .ajax call that grabs the data from a url and appends it to the div by calling div.html(data) inside the .ajax callback, so it's essentially loading a PageB by getting the data (content) and appending it to the div that's calling the dialog("open")...nothing special here I don't think.
My question: In Page B, how do I reference the dialog so I can do some things to it? For instance in Page B's content that I received back from that .ajax call and added to the div via .html(data), there is a button and when clicked I need to close the dialog.Right now my buttons are not working inside the dialog because one of them closes out the dialog and the other should redirect to a new page but both do not work now because I have no reference to the dialog that it's in to manipulate it. So I need reference so I can close the dialog via some jQuery that will reside in PageB (data).
i have server 2008 and real ip i need to publish the website on this server to allow external access for people what are the steps to host the website on iis7 under windows server 2008 should i configure dns first?
When we include Javascript files in our aspx files. either we write all the Javascript code in the Head section of aspx or we link to an external Javascript file.
So when the clinet makes a call to that page. Does that mean that all the Javascript will be loaded on the client side?
If that is the case then does not that mean that it will slow down the loading of page as all the Javascript has to be loaded on the client machine?
I have an asp.net web page to serve large file downloads to users.
The page is hosted on IIS7, Windows Server 2008.
The strange thing is that users can download at good speeds (2MB/s) when I don't add a content-length response header but as soon as I add this header, download speed drops to somewhere around 35kbps/s.
This is the code:
[code]...
Of course I can leave the content-length out but the user will not know how big the file is and how long the download will take...which is annoying.
I am loading an external file 'contact_us.aspx' into a jquery dialog box. the external page contains a form.When the submit button is pressed it closes the dialog box and changes the page to contact_us.aspx. is my code correct or is there a different way of doing this. see my code below, thanks.
When I undo an external dll in TFS, it looks as though it worked because I can compile with no errors. However, if I look in the folder where the dll lives it still shows the newest dll, which technically would not compile if I just did an Undo.
So, my question is, where does the old dll live after I do an Undo?
I am using some webrefernces for reporting services. The first time they load they are really really slow. Is there any way to reference the files locally?
I have a Class Library Project that I'm using in another Web application project. I've noticed that this has just copied the Class Library DLL into the Bin folder of the Web Application, and also that if I make some changes to the Class Library and rebuild the project, I also have to rebuild the Web Application for the changes to take effect there.
This kind of defeats the object why I'm creating a Class Library as I'm hoping to put some generic functionality in there that will be available to a number of different web applications.
I am used to working on C# .NET windows applications. I am interested in starting with web applications. good ebooks or references for C# ASP.net to begin with for a guy like me who is comfortable with windows applications?
I have an Asp.Net TextBox control whose enabled property I need to evaluate using a separate (from the code-behind of the containing page) class. The way I want to do this is:
The reason I need to pass a reference to the calling control is to see if there are other properties I may need to consider before disabling it. The 'this' keyword, however, refers to the page as a whole. And since my TextBox can be within a FormView, or GridView, I can't easily just get the control byexamining 'page.myTB'. I don't want to have to start ploughing through GridViews on the receiving class file just to retrieve a reference to the TextBox.Does anyone know how to refer to the control inline in this way?EDIT 1: This will need to be done to several different types of server control, such as ListBox, CheckBox, RadioButtonList, etc.EDIT 2: The best I've come up with so far is:
<asp:TextBox ID="myTB" runat="server" Enabled="<% ToggleControl(myGridView.FindControl("myTB") as WebControl, "someBusinessValue") %>"
I have two applications(A,B) both are developed in .NET3.5 and used LDAP services.Application-A in one screen we are fetching users with role "project Leads" and it is taking 10 min time
Application-B in login screen we are checking user is authenticated or not and it is taking 20 sec
Both the applications are using the same service then way there is a time span problem. Please suggest me what can I do to improve performance of application A?
I am using Reporting Services to render a report directly to PDF. It requires that I use two web references: ReportExecution2005.asmx and ReportService2005.asmx. The performance on web references seems really poor. Since my web server (IIS7) and my SQL Server (2008) are on the same box, is there a way I can reference them directly? If not is there any way I can explicitly cache them or something. First load is really really slow, second load is perfectly acceptable.
I have a web application that dynamically loads assemblies based on database configuration entries to perform certain actions (dynamic plugin style architecture).The calls to the objects are in a Factory Pattern implementation, and the object is cached (in a static dictionary<> within the Factory) as the calls can be made many thousands of times in a minute.The calls to this factory are made from both the main web application and a number of webservices, some in different assemblies/projects.When I need to update one of these DLLs, I have to recycle IIS to get the DLL released.As this has an impact on another application on the server, I wanted to know if there was a way I could release the DLL without restarting IIS?
I have a separate environment for development. On my production server, where we mostly host web applications, we deploy the precompiled version of the web project. While doing development I work on three projects. One is the WCF Service. Two, is the class library project. This library has classes which makes calls on the WCF service. And the third, is a web project which consumes the class library.
The last two projects come under one VS solution. I host the wcf service in the development environment. This service is added as service reference to my class library. I am unable to visualize what will happen once I add the DLL (viz output of the class library project) to my web project, and, the web project has to be deployed. At the time of deployment I have to change the service reference (the url of the svc file will change as it has to point to the production wcf service). How to go about this? What must I take care when adding the dll to the web project?
how to create a snk for all existing dlls(multiple) in n-tier ASP.net application?I have created a asp.net application using n-tier. My web layer contain refrences of all layers(data,facade,core,common).but when I try to create strong key of web layer it throws error as "refrences assembliy can not have Strong name".
why is it that when i debug my application in chrome it is relatively slow between postbacks compared to internet explorer? when i use internet explorer, its lightning fast.
I have a strange situation on a production server. Connection for asp.net get queued but the CPU is only at 40%. Also the database runs fine at 30% CPU.
Some more history as requested in the comments:
In the peak hours the sites gets around 20,000 visitors an hour. The site is an asp.net webforms application with a lot of AJAX/POSTs The site uses a lot of User generated content We measure the performance of the site with a testpage which does hit the database and the webservices used by the site. This page get served within a second on normal load. Whe define the application as slow when the request takes more than 4 seconds. From the measurements we can see that the connectiontime is fast, but the processing time is large. We can't pinpoint the slowresponse the a single request, the site runs fine during normal hours but gets slow during peak hours We had a problem that the site was CPU bound (aka running at 100%), we fixed that We also had problems with exceptions maken the appdomain restart, we fixed that do During peak hours I take a look at the asp.net performance counters. We can see behaviour that we have 600 current connections with 500 queued connections. At peak times the CPU is around 40% (which makes me the think that it is not CPU bound) Physical memory is around 60% used At peak times the DatabaseServer CPU is around 30% (which makes me think it is not Database bound)
My conclusion is that something else is stopping the server from handling the requests faster. Possible suspects:
Deadlocks (!syncblk only gives one lock) Disk I/O (checked via sysinternals procesexplorer: 3.5 mB/s) Garbage collection (10~15% during peaks) Network I/O (connect time still low)
To find out what the proces is doing I created to minidumps.
I managed to create two MemoryDumps 20 seconds apart. This is the output of the first:
!threadpool CPU utilization 6% Worker Thread: Total: 95 Running: 72 Idle: 23 MaxLimit: 200 MinLimit: 100 Work Request in Queue: 1 Number of Timers: 64
and the output of the second:
!threadpool CPU utilization 9% Worker Thread: Total: 111 Running: 111 Idle: 0 MaxLimit: 200 MinLimit: 100 Work Request in Queue: 1589
I am developing sample web application which use C#, and LINQ to SQL. The application is very very slow and it takes about 2 secs to navigate between pages. I have already used SQL Profiler and monitor the calls to the SQL server. All SQLs looks normal to me and their execution time is always about 1 or 2 millisecs.I tried the same application in another PC and it is the same. So, it is not because of my computer performance.
My application runs quickly when debugging locally, but slows down to an almost unusable state once published. Oddly, the published version seems to be grabbing data very quickly from MSSQL - it's the part where I'm updating datatables and gridviews that is very slow. For example:
I know that when a .NET web application first starts, it will be slow on first access. For that reason, I created a 'keep alive' page in the application that is requested every 8 minutes or so by a scheduled process. I hoped that that would be sufficient to prevent the application from 'unloading' and thus suffer from this initial long startup time again and again.
I can see in my logs that the 'keep alive' page is indeed triggered every 8 minutes, 24 hours a day. But when I actively start working in the application after some time, eg. the following day, I still have to wait a while for it to start up. Once the first page is loaded, all is fine again.
There's a web app I've been assigned to, which is running very slow. It is a site that sells products so it is database driven, however even pages that do not query the database are loading very slow. The pages use master pages, and the code is in VB.NETI checked with fiddler and the time it takes to load basic (non database driven) pages are about 5.5 seconds on average.
What are some tools that can help me determine the cause of the slow speeds, and any recommendations as to how to speed it up, or potential issues that could cause it?UpdateSo I messed around with the code piece by piece as I wasn't getting anywhere with these tools. As soon as I remove master pages, and I include the same code that's on the master pages in the .aspx page itself, the speed improves drastically (approximately 5 times faster load times).What might cause the master pages to cause load times to slow down so much?