Web Forms :: Size Of Folder Affect The Performance Of Application?
Mar 17, 2010
I am developing a site similar to forums. I allow users to Post articles and upload files regarding the articles. I am not storing the Attachments contents in DB instead i store only the path of the attachment file and store the actual file in physical folder in the Project directory. WHat i want to knw is if the size of the physical folder increase will it affect the performance of my application.
I have a dataset that comprises of multiple tables (about 20) linked together with constraints to reflect the relational database. My question(s):
1) Will having multiple tables affect performance? I'm asking this as there are often times I would not need to get data for all the tables, but only for specific tables.
2) If it does affect performance, would breaking up the tables into their own datasets be the best solution. It seems to me that by doing that, you would lose the ability to apply constraints among tables and tables that might exist in multiple datasets will be subjected to replication.
I would like to have everything in one dataset as I get a view of the related tables and their relationships and also because the Fill method allow me to fill tables, maintaining the hierarchy of the data, but performance is still a key factor.
Would adding "X-UA-Compatible: IE=EmulateIE7" as an http header to trigger IE9 to render in compatibility mode negatively affect performance on the server? I made the change today, adding it, but now some areas of our site are behaving sluggishly, though we did have a large number of changes come in overnight. I know that we can't single out this as the problem, but my boss seems to think it is.
I am not sure if this is the right forum. I can not find a forum for LINQ.
I am working on an application using LINQ. Application performance is not up to par and my tests show that it is LINQ queries that are slow. I was wondering if anybody can recommend where I can find an article about optimizing LINQ performance maybe by compilation or other methods.
In my website for one project we have 5000+ records in sql database. If I give page size is 100 for gridview then the paze size showing 800kb in firebug net panel and it is taking 20-30 seconds to load webpage. If I give page size is 10 for gridview then the paze size showing 160kb in firebug net panel and it is taking 5-6 seconds to load webpage. If I give page size is 5for gridview then the paze size showing 130kb in firebug net panel and it is taking less than 6 seconds to load webpage. I noticed that due to heavy paze size the performance decreasing but as per my requirent the paze size should be 100 or above and the page should be load less than 2 seconds.
I have uploaded a .jpg image and resized as Small and Large thumbnail. The small image is (135 X 173) but after resizing the small thumbnail image file size is of 40.8 KB. And mostly all images (around 64) thumbnails having file size more than 40 KB. Due to which thumbnails taking lot of time to get display.
Is there a way through which we can reduce the image file size upto 8 to 12 kb, but without effecting image quaity??
Below is the source code that i am using for resizing an image:
If FileUpload1.HasFile Then Dim stream As Stream = FileUpload1.PostedFile.InputStream Dim x As String = IO.Path.GetExtension(FileUpload1.FileName).ToLower If x = ".jpg" Or x = ".jpeg" Then Dim image As New Bitmap(stream)
I am working on AJAX-ASP.NET site. This site is taking 25-30 sec to load the contents. I want to improve this site performance. Is there any settings required in IIS? How to check IIS log files?
As an extensibility option for my web application I would like to allow the user to add sections to it by simply making new folders and uploading files to them. So, for instance, if they created (in the webroot): /UserContent/StaticPages/First/Second/index.txt. Then in the main menu there would be a new menu point "First" with the submenu point "Second", which would lead to markdown-parsed version of index.txt. This way the user can upload any static menu points he wants (such as "About us", "Contact us", "Our Mission" etc.) Another extensibility point is that I would like them to add headers/footers to specific webpages in a similar fashion. I'm using ASP.NET MVC3, so there are pretty URLs, and it would be easy for the user to just create: /UserContent/Additions/Store/Categories/35/header.txt
And when someone opened [URL] the markdown-parsed header.txt would be prepended. This is simple for the user, and simple for me (I don't need to make sophisticated admin panels and WYSIWYG editors). However I'm worried about the performance. This scheme means that on each GET request I would need to scan the /UserContent/StaticPages folder and check for the existences of the header/footer files (several, because parent levels can have their header.txt/footer.txt as well).
I could cache the result, but then I have to manage the cache and the user will need to know that changes can take up to X minutes to display. Would this be a premature optimization? There won't be much data in those folders, so Windows will probably be able to easily cache their contents itself.
I know this is a bit like how long is a piece of string question, but I'm wondering how big some peoples ASP.NET Temporary Files folder, gets. I wish to move this folder to a RAM DISK so I'm trying to see how big this should be set, to.
I'm trying to calculate folder size of remote server.I have credentials of server.By using WMI,I could fetch the existing Logical drives in remote server. But,couldn't get Folder name or details..Here,my folder is not shared one..I like to get WMI solution for it..if it not possible,atleast any dos commands to get file size of remote machine like this, \ServerIP:username@password\foldername -dir/s . .
I have a strange situation on a production server. Connection for asp.net get queued but the CPU is only at 40%. Also the database runs fine at 30% CPU.
Some more history as requested in the comments:
In the peak hours the sites gets around 20,000 visitors an hour. The site is an asp.net webforms application with a lot of AJAX/POSTs The site uses a lot of User generated content We measure the performance of the site with a testpage which does hit the database and the webservices used by the site. This page get served within a second on normal load. Whe define the application as slow when the request takes more than 4 seconds. From the measurements we can see that the connectiontime is fast, but the processing time is large. We can't pinpoint the slowresponse the a single request, the site runs fine during normal hours but gets slow during peak hours We had a problem that the site was CPU bound (aka running at 100%), we fixed that We also had problems with exceptions maken the appdomain restart, we fixed that do During peak hours I take a look at the asp.net performance counters. We can see behaviour that we have 600 current connections with 500 queued connections. At peak times the CPU is around 40% (which makes me the think that it is not CPU bound) Physical memory is around 60% used At peak times the DatabaseServer CPU is around 30% (which makes me think it is not Database bound)
My conclusion is that something else is stopping the server from handling the requests faster. Possible suspects:
Deadlocks (!syncblk only gives one lock) Disk I/O (checked via sysinternals procesexplorer: 3.5 mB/s) Garbage collection (10~15% during peaks) Network I/O (connect time still low)
To find out what the proces is doing I created to minidumps.
I managed to create two MemoryDumps 20 seconds apart. This is the output of the first:
!threadpool CPU utilization 6% Worker Thread: Total: 95 Running: 72 Idle: 23 MaxLimit: 200 MinLimit: 100 Work Request in Queue: 1 Number of Timers: 64
and the output of the second:
!threadpool CPU utilization 9% Worker Thread: Total: 111 Running: 111 Idle: 0 MaxLimit: 200 MinLimit: 100 Work Request in Queue: 1589
Web forms .NET application, is starting to meet 64 bits Windows servers and we have a 32bit DLL from one of our component vendors (its a hardware key, actually). I checked the instructions here "compiling a .net application with either a 32-bit or 64-bit dll". It is also what our vendor sugested. question: is there a performance impact on compiling a web application to run on x86 mode ?Please don't flame me for using a hardware key
I have a ASP.NET web application (.NET 2008) using MS SQL server 2005, I want to increase the performance of the web site, If anyone have an article contains steps to do that, step by step , In SQL(Indexes, ..... etc.) and in the code.
There's a web app I've been assigned to, which is running very slow. It is a site that sells products so it is database driven, however even pages that do not query the database are loading very slow. The pages use master pages, and the code is in VB.NETI checked with fiddler and the time it takes to load basic (non database driven) pages are about 5.5 seconds on average.
What are some tools that can help me determine the cause of the slow speeds, and any recommendations as to how to speed it up, or potential issues that could cause it?UpdateSo I messed around with the code piece by piece as I wasn't getting anywhere with these tools. As soon as I remove master pages, and I include the same code that's on the master pages in the .aspx page itself, the speed improves drastically (approximately 5 times faster load times).What might cause the master pages to cause load times to slow down so much?
I want to test asp.net application for stress, load and performance. Please guide me what free tools should be used for this ? Is there any tool which particularly can be used to analyse individual components, functions and scripts and resources they use ?