DataSource Controls :: Reading Excel And Flat Files From Application
Mar 16, 2010
iam Working on Asp.net Web apliation using c#.net. in system folder (Ex: E://Santhu) i have few Flat files and Excels Files .So My Application should read those files automatically and should store the data into my database..
iam Working on Asp.net Web apliation using c#.net, in system folder (Ex: E://Santhu) i have few Flat files and Excels Files .So My Application should read those files automatically and should store the data into my database..
i've got some problem reading excel files to store into sqlserver database. currently my code has a fileupload control and a button for me to save the excel file into a folder and read the data inside the excel file and write it into the sql database. The code works only if the excel sheet name is Sheet1.xlsx, however i tried upload a diff file named ImportUserFile.xlsx and it gave me the exception below. 'ImportUserFile$' is not a valid name. Make sure that it does not include invalid characters or punctuation and that it is not too long. i have also tried other file names and it seemed only Sheet1.xlsx works.
my problem is my asp.net and C#.net application should read flat files(.txt,.doc) from folder automatically and read the data from file and stored into database
my file like this :
BNK20100312APPHU6658994 // BNK,20100312(date),Ap,PHU6658994 So we have to store these.. BNK20100312APPRU5104220 BNK20100312APPRU5761800 BNK20100312APPZU3739288 BNK20100312APPZU4454037 BNK20100312APESU2552894 BNK20100312APAXU4302543 BNK20100312APCKU1613369
i have flat files in some direcotry. each flat file contains around 60,000 rows of data. earlier we were using excel sheets to upload data. but now the challenge is it is taking more time to upload data into DB tables. which is the best way to upload such a huge data into db tables. can we use SSIS for uploading data. if so, how can we achieve that.
I have a flat file with text that has a special character like this: "RIBAŽ Generic Strip Labeling." so the registerd symbol is coming as square when i am simply reading the file in c#. is there a way to read the symbol as is and print on screen? i am working in on a console application. Also other german characters, not all but a few are showing as squares as well.
I'm trying to read a CSV file into a DataTable using OleDbConnection and OleDbDataAdapter. It works loading the DataTable, but only does 255 columns. The file has 362 columns.
Want to create a web based application for reading and writing excel files.
Issue :
1. Want to upload an excel file and store it some location. 2. Read data from excel file, performing some sorting and filtering 3. Write filtered data to new excel file.
I want to send the email through asp.net where "To " email address list store in excel file. Send email to all email Id store in excel file. How to send the email reading all email id from excel files.
I am working on an SSIS package that needs to read a flat file and load the data into a SQL table. The problem is, the flat file name changes because it incorporates a date similar to this:
fileName_2010-04-12.txt
I found this site that shows how to set up a dynamic flat file connection:
[URL]
I understand the principle, but what I'm struggling with is using it in my data flow task. I need to configure the column mappings. In the PROPERTIES of the flat file connection, under expressions, I set up a ConnectionString with the following:
@[User::DataLoadDir] + @[User::priceFileName]
where [User::DataLoadDir] is the path where the file resides and [User::priceFileName] is a variable that should have the name of the specific file I want to read.
However, when I try to set the column mappings in the flat file connection, it's pulling the column headings for a different file in the same directory.
I have to get input from user in flat file format.then convert it to xml. Then, i have to store the entries into a datatable in the database. say 1,2,3 a,b,c be the user entries. it shall be stored into a table with 2 columns.how to do this?
I need to read an Excel work book with 2 sheet which have more than 60,000 records. The application is an ASP.Net application so the performance matters.
Which approach should I take? Sohuld I do it using Microsoft.Office.Interop.Excel or should I do it using OLEDB in ADO.Net?
I can directly use to read and display Excel or CSV files using asp web control(GridView, sqldatasource etc). I do NOT want to upload the file first to an sql database before reading them, I just want to read the files directly. I am using asp.net 2.0 and c#.
I don't know how to read data into variables from a SELECT query. I can add, update, and delete queries with no issues. I have browsed a few books, but everyone uses a different approach. Suppose I want to to run a query and 1 record is returned. How do I get the values into my variables.
Suppose my table name is "Job Orders", and the row returned has these 3 fields:
I have a C#/Asp.net web application which caches SQL DB schema information. Is there a way that I can determine what child relationships a given table has? I have looked into using the DataTable.ChildRelations property, but it is never populated with items unless i manually create a relationship using the .add property. I also have access to DataColumn information for each column in the database.
I need to populate a database (SQL Server) with some data from a templated excel document.Here are some approaches:
A) User upload a file.xls or file.csv that the server reads, checks and updates the database (cons:the uploaded file might be too big. file.* needs to be in correct format)
B) User downloads a tool.exe to read and upload file.xls (cons: maintenance of the tool.exe)
C) Use of a web service in excel? (don't know how to do it) but this article might give you a hint
Which one is the best to use? Are there any others?
In my application, i have one fileupload control which will take excel file. I want to read this excel file and want to store in datatable. How will i do this. I am using visual studio 2008.
I'm trying to store multiple selected checkboxes from a single checkboxlist, but i don't know where to start. Imagine i have three tables: tbl_events, tbl_organisors and tbl_settings and i want to store the organisors of an event. Each event can have multiple organisors, so each organisor should be added as a single row into tbl_settings. I know how to bind tbl_organisors to a checkboxlist, but how do i store the results (meaning: add the rows to tbl_settings) ?
tbl_events: Event_ID (PK) and Event_Name tbl_organisors: User_ID and user_name. tbl_settings: Settings_ID (PK), Event_ID and User_ID
Checkboxlist:
[Code]....
SQLDatasource:
[Code]....
Initially all checkboxes will be unchecked, but users must be able to select multiple organisors from the checkboxlist and then store them into the DB by submitting a form. So the query will look something like: INSERT INTO tbl_settings (user_id, event_id) VALUES (@user_id, @event_id). And this is where my head cracks
1) How do i loop through all selected checkboxes and store each value (user_id) in a new row in tbl_settings? 2) After saving, the users reloads the page and the organisors he just selected must be selected by default in the checkboxlist. How?
The short story is, I'm developing an SSIS package to select from a remote Oracle database (read: no Windows Authentication) and store into an SQL Server db. I can execute the package just fine in development and after importing, but when I run it as a scheduled job, I get an ora-01017 error (bad un/pw). The details are, I built this package and tested in in BIDS but came across the problem that BIDS really doesn't save the password I'd set for the Oracle connection. After poking around a bit, I found this
[URL] which showed how to create an XML configuration which the package could then read. Fiddled around with that, got it working in BIDS. Connected to Integration Services on the server on which I want to run the package, imported it, right-clicked and did the "Run Package" option as a test, it works great. However, when I connect to the same server's Database Engine to create and schedule a job, I cannot connect to Oracle. For some reason, it seems as if the package is not reading the configuration when run in schedule.
I've tried removing the job and re-working it, thinking maybe it didn't pick up the new version when I re-imported the package. A new job had the same problem. Next, I tried re-naming the .dtsConfig and doing the "Run Package" option as a test. Failed with the same error. A bit surprised, I thought I'd get some kind of file error and we wouldn't get as far as trying to talk to Oracle. Anyway, thinking maybe it's permissions on the configuration, fixed the .dtsConfig name and gave full access to Everyone. That ought to let the Agent see it. Scheduled up the job, still can't talk to Oracle. So it seems as if the scheduled version is ignoring the configuration. Anyone have any idea what I'm missing, that scheduling an SSIS package makes it act differently than running it manually? BTW, the instance under which we're running the package is VS 2005. BIDS 2005 to build the package.
I'm Using a DataReader to Read about 100,000 records, my code look like :
[Code]....
I have a per row computation in reading the rows on GetMyType2Info() method, and this method also another method that I return only one record by ExecuteReader(). the problem is that when the MyTypeList (List<T>) has about 21,000 Items of MyType (<T>) System.OutOfMemoryException is thrown. wanted to know how to resolve this and wanted to know more about the maximum size of List<T> capacity. in addition the execution of the main Query with 100,000 records take about 4 seconds in SQLServer a record has only types of nvarchar , float,datetime and int