I am using PL/SQL Developer.I have a below query select distinct
DATA_SRC_CD as Source, FILE_NM as File_Name, PATH as File_Path from PKG_SFI_REPORTS.DATALOAD2SUMMARYETLRPRT("FACTS","CM_DDL_PARTY_2011_01_12_08_24_09.DAT") In the above Query PKG_SFI_REPORTS => Package, DATALOAD2SUMMARYETLRPRT=> Procedure. "FACTS" and "CM_DDL_PARTY_2011_01_12_08_24_09.DAT" => Parameters Now I need to know the Data Types of "DATA_SRC_CD" and "FILE_NM" and "PATH". How can I Know?
I have a code potion that runs a particular SQL and Fill the data in to a DataTable:
This code is working perfectly for a Query in my local machine without any error. But when I am trying to run this code in my friends machine (which is in the same network) I am getting the error 'ORA-00932: inconsistent datatypes'.I have tried to run the SQL in the database also (with the same user), but got no errors there. Could somebody please help me to get out of this problem?Here is the SQL I am running, the query is basically returning all the jobs that are being scheduled under Oracle 10g scheduler and the count of errors they have from the previous day:
On one of our tables, a default clustered index is defined on the primary key by sql server 2005. The table also contains forign key columns and has got over 2 million rows. There are very frequent INSERT, UPDATE and DELETE functions on this table. Recently some of the SELECT queries which primary operate on the forign keys have become very slow (resulting in application timeout). To fix the problem, I defined an index on one of the primary keys which gave me very quick results. I don't know much about indexes.
1. Since I defined a new index (on the forign key) in addition to the existing default index, what kind of index (clustered or non-clustered) have I defined?
2. What performance impact does it have on the rest of the system? Is it going to break something else?
3. I read about fill factor. What is this and how does it effect the system?
4. Is the new index physically going to make any changes?
5. In case I delete this index later on, will the database return to the existing state or it will be a different state from now?
I'm creating a table in an asp.net code behind dynamically, and I want to have a footer row that only has 2 cells. The first should span all the columns in the table-1. Is there some way other then keeping track of the # of columns in the table manually for me to set the colspan to be # of all the columns in the table-1? Preferably a HTML or CSS solution?
I've the following code which successfully makes an average for all the columns from a table. What I need to do though is ignore certain columns in this equation.
Dim totalNumber as Double = 0 Dim count as Integer = 0 For x = 0 To xyz123.Tables(0).Columns.Count - 1 Dim current as Double = 0 If Double.TryParse(xyz123.Tables(0).Rows(0)(x).ToString(), current) AndAlso current <> 0 Then count += 1 totalNumber += current End If Next Dim averageRating as Double = totalNumber / count
I'm writing a reporting tool that, as a starting point, needs to create a DB table from a csv file. This csv file will have been created from exporting data from our old system (with the new system being the reporting tool) and can have up to 300 columns of data. I've been trying as follows:
but am finding that only the first 255 columns get put in. I am assumg that the Jet driver is the one that is limiting the columns, or is there a registry setting somewhere that'll solve it quickly?
Is there another approach that'll work? I've though of the following, but as it'll take a while to recode and test each time I thought I'd ask the experts first!!
BULK INSERT: would mean that I'd need to pre-create the table first, but if the issue is with the Jet driver I'd surely get the same problem? Use a datatable: it seems pretty fiddly and probably a lot slower, but I guess I could read into a dataset and then bind that to a table to insert? Again the table would need to exist and the driver that I use to read from the csv would be the issue? Do it one row at a time in a loop in the code: assuming I can read the file as text, I could then split the data in code into an array and create a insert statement (again to an existing table) for each row. With what could be 1000s of rows I can't imagine that this is the fastest solution!
i done one web application in this application i need to transfer the data from one table to another automatically when the time is 10.00PM. i need to write this function sepearte time function not in page load function.the user select any of page the data should be transfer from one to another without page load only using time control function..
I am building a little app to upload a picture and various downloads for different file types for the photos, but I also need be able to select the catergory they will go into. So I have two tables one that stores the catergory names and one that holds the paths to the photos and downloads the tables are related through the catergoryID.
What I need is those catergories to show up in the photo's and downloads table so that when I go to upload the photos I will be able to select the categories.
I have also built the DataSet and the Business Logic.
i did the mysql data export into excel in row wise when button click... here i want the same data export into columns wise..i used datatable and dataset for this function.. here i need all data should export in column..here i used some code..
I have a page within our site that allows admin users the ability to upload configuration settings into the system instead of submitting 1 at a time.
While testing yesterday, i found that one of the upload pages allowed me to upload a excel file that had alot more columns than expected and when i checked the database, the first few columns that were expected took the values from the file, and everything else was ignored, but wanted to see if there is any way to check the number of columns in the file to make sure it matches the upload logic before accepting it.
True) to update database row that returns null values from date and numeric columns,Whenever i try to update same row with null (or empty) values on both date and numeric columns, it results in sending '01-Jan-0001' for date column and 0 for numeric column. I want the user to have the options of null or corresponding values for the update.
I have an Excel file called Products.xls .I have Columns A and B, with the titles NAME and QUANTITY.The name of the sheet is SHEET1.The file has about 40 lines.How do i show these data on a Gridview or Listview ?
I'm working on locking down some MySQL user accounts. At the moment I typically grant my user accounts execute privileges over the required stored procedures within a schema.Now I'm looking at the proc table in the mysql schema. The common wisdom that's quoted many times on various websites says "Stored procedures require the presence of the proc table in the mysql schema".By granting select access over the proc table stored procedures will work. But if I removed the above select privilege and grant the user execute privilege over the entire mysql schema the procedure will also work.
Does anyone have any ideas about the security issues by choosing one of the above over another?I would prefer to lock the proc table down all together so the user cannot see it cannot select from it.
If I use Microsoft.Jet.OLEDB.4.0 to retreive for example named range or sheet from Excel and use it as data source for some server control on the ASP page, everything works fine. But when I try to get table from Excel like this: "select * from [Table1]", debugger complains that database engine could not find object Table1. So how should I reference Table1 in Excel file so that the database engine could find it?