DataSource Controls :: Dealing With Large DBML Files?
Jan 4, 2010
I used SQLMetal to generate a DBML file, which worked fine. The problem is that since this DBML file is so big (185,000 lines at 840Kb) it takes literally hours to load in VS2008. What is the recommended way to resolve this? I tried using a third party tool call SqlMetalInclude to break the DBML up into multiple data contexts, but that tool generates only designer.cs files, not dbml files.
I want to upload some large files from a web page to MS SQL Sever database, I am very sure that it is sounding weird.
File sizes are around 100MB.
I am having following settings,
SessionTimeOut period = 60 Mins,
Server Operation timeout = 60 Mins,
SQL Connection Timeout = 4 mins (Not sure if this is helping)
This is page is going to be used by our client only once a week & as they are having web farm environment we are avpiding to store these files on file systems.
Currenty we are able to upload files with sizes upto 8 MB succesfully. But when we are uploading a file of 100MB it fails, its for sure that operation takes lots of time.
If I make changes to a table or view in the datacontent (dbml), is there any way to refresh the fields in the dbml without deleting the datasource and inserting it back again?
I have MyDB.dbml in my Models folder that was created by dragging tables from Server Explorer. I'd like to make some changes to the design of those tables. Before I do that, how do I get the dbml to refresh itself?
I do expect to have to fix code where field definitions change & that's okay. I just don't want to have to start over from scratch.
I have a dbml file which I have added a couple of tables to. The database also contains synonyms which point to other tables in a different database. I can't seem to drag the synonyms to the design surface though?
i have some confusion using compund from in linq.I know the concept of compound from but in database table how can we use.can anyone give me example of it.
Is there a quick and easy way to update my LINQ dbml files?Right now, if I update a Stored Procedure, I need to open the dbml designer file, delete the old SP, save, drag and drop the new SP, and save again. Dont' get me wrong, this isn't "hard"... just with all the fancy new technology out there, it would be nice to be able to just right-click and hit "update".
We are going to develop a website where we need to store large number of files. The file size may be uptp 50 MB. What approach should we follow:
1) We should store files in Database 2) We should have a directory and store all the files in that 3) We should hire a SAN Storage and use this seperate location to store the files.
Is there a way of filtering large CSS files for the only required selectors on a page, and creating css files that contain just these selectors?
Case: I have a very large CSS file that I want to filter on a per page basis, so that the file size is cut down and can be cached by mobile devices. I was thinking along the lines of something like a server side dust me selectors tool.The particular project I am working on is using ASP.NET MVC.
I want to insert a large amount( approx 7000 records) of data into a table, My scenarios is that i need to check whether the record already exists then it will not insert it. Currently i am using a stored procedure for this which gets called for each record.
IF EXISTS(SELECT 'True' FROM MYTABLE WHERE ID = @ID) BEGIN
--This means it exists, return it to ASP and tell us
SELECT 'This record already exists!' END ELSE BEGIN
--This means the record isn't in there already, let's go ahead and add it
SELECT 'Record Added' IF EXISTS(SELECT 'True' FROM [Aadil].[dbo].[SharePointSitesData] WHERE ID <> @ID)
I need to populate a database (SQL Server) with some data from a templated excel document.Here are some approaches:
A) User upload a file.xls or file.csv that the server reads, checks and updates the database (cons:the uploaded file might be too big. file.* needs to be in correct format)
B) User downloads a tool.exe to read and upload file.xls (cons: maintenance of the tool.exe)
C) Use of a web service in excel? (don't know how to do it) but this article might give you a hint
Which one is the best to use? Are there any others?
If I have a huge Order-table and I want to display records from it using paging, what is the best way to select the records from the database?
If I first do like this:
[Code]....
..and then select the records I want based on RowNum it seems like I select ALL records in the first query anyway. I guess I want to do something like SELECT TOP @PageSize FROM Orders...
We use the MojoPortal to a website and have some problems to upload files that is around 100 MB with the upload module. (Pleas note that this has probably nothing to do with MojoPortal but with the ASP.NET and the IIS)
The MojoPortal is set to use regular file Upload(not Neat Uploader) and to be able to upload big files we have set the following :
The problem is that the upload will cacel after a couple of minuts (Aborted).
Is there any other values that I need to set to make this possible? The MojoPortal itself should not have any settings for this as far as I know so its regular ASP.NET 4.0.
Anyone got some good pointers at an open source (article for creating your own would even be better) component to upload large files.SlickUpload for instance works great, and surely worth the money, but as this is for a pet project, a paid solution is just not what I'm after.
I am building a website where i need a page where user can upload large video files, i have created WCF service with streaming but i am calling that WCF service from Button_Click event of web page.
I have used below mentioned article for WCF service creation
WCF Streaming
I have used streaming as it should be efficient and should not be buffered in memory of server.
Now questions
1) I am having doubts that the entire file is uploaded to the web server and then it is transferred to WCF Service server...if this is true then i am not getting advantage of streaming as well as iis and web server will be down very soon if user uploads large file or multiple user are uploading files con currently
2) Is there any other efficient way to do same operation with some other technique
EDIT :
If I am not calling WCF Service method from ASP .Net code in that case also it is transferring bytes to the web server which i have checked with HTTPFox
I have checked above thing with upload control and putting one button on UI whose click event is bound to one method in code behind.
So, still i am having that confusion that how data is transferred
Client Machine - Web Server (ASP .Net Application) - Service Server (WCF Service) Client Machine - Service Server (WCF Service)
NOTE : If i am putting a debug point on button_click and uploading 10 kb file it hits that in less then 1 sec. but if i am uploading 50 mb file then it is taking time.
I placed code of calling WCF service inside that button_click event
I'm facing a problem in our current running application. There is a table in database which has 67 columns,among these 18 columns are nvarchar(max) datatype and contains 30K records. Problem is now that SELECT * FROM myTable it takes around 3 to 4 seconds. Which is degrading our application's performance.
I have contact mails large than 700.000 Mails and i want to insert them in sqlserver using c# in a few minutes not hours i know that pre impossible but , i know there`s way to do it
I need to create an upload site to upload large files over 2GB I want ot create a site like [URL]. Once these files get upload i want them to have a link to the file created but the link encrypted. I know there is a limit to http upload. I have used a bunch of the flash upload web apps but are capped at a specfic mb becuase of .net. What options are out there.
I'm using a web form that allows users to upload media files. The code works great on small to medium size files, but I've found that if a file is really big(like bigger than 15MB), the user will get a 404 error. Currently I'm using the code below to handle the file. Does .NET provide another way to handle larger files?
I used this sample to work around the issue we were having with large files. [URL] Unfortunately, when I attempt to download large files of 30MBs or more, the download times out and the user gets a partial download. It doesn't seem to be a consistant percentage of the download either. I attempted to download a 50MB file and got to 33MB. When trying a 30MB file, I downloaded 24MB.Below is my code.
if (File.Exists(strFilePath)) { fileName = System.IO.Path.GetFileName(strFilePath); Response.Clear(); system.IO.Stream iStream = null; byte[] buffer = new Byte[10000]; int length; long dataToRead; try { iStream = new System.IO.FileStream(strFilePath, System.IO.FileMode.Open, System.IO.FileAccess.Read, System.IO.FileShare.Read); dataToRead = iStream.Length; //FileInfo file = new FileInfo(strFilePath); Response.ContentType = "application/octet-stream"; Response.AddHeader("Content-Length", iStream.Length.ToString()); Response.AddHeader("Content-Disposition", "attachment; filename=" + fileName.Replace(" ", string.Empty)); while (dataToRead > 0) { if (Response.IsClientConnected) { length = iStream.Read(buffer, 0, 10000);..................
I'm testing a very simple aspx page on Visual Studio's own ASP.NET Development Server(the local server). On the webpage there is a FileUpload control which can upload jpg file up to 2MB without problems. On uploading bigger files, the browser immidiately show "The web page cannot be displayed". It does not show any exception which really puzzles me. "The web page cannot be displayed" is normally caused by network problem, but in this case it's a local server and it can handle smaller jpg file fine. Whta's the problem here?